var/home/core/zuul-output/0000755000175000017500000000000015067773734014550 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015070007170015466 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005170147215070007161017701 0ustar rootrootOct 03 16:24:31 crc systemd[1]: Starting Kubernetes Kubelet... Oct 03 16:24:31 crc restorecon[4689]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:31 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:32 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 16:24:33 crc restorecon[4689]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 16:24:33 crc restorecon[4689]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 03 16:24:34 crc kubenswrapper[4744]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 03 16:24:34 crc kubenswrapper[4744]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 03 16:24:34 crc kubenswrapper[4744]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 03 16:24:34 crc kubenswrapper[4744]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 03 16:24:34 crc kubenswrapper[4744]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 03 16:24:34 crc kubenswrapper[4744]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.641097 4744 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.643942 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.643959 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.643964 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.643968 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.643972 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.643976 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.643980 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.643985 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.643991 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.643995 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644000 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644044 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644049 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644055 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644060 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644063 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644067 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644070 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644074 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644078 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644082 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644085 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644089 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644093 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644096 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644100 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644104 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644108 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644111 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644115 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644119 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644123 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644126 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644130 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644134 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644137 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644141 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644146 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644149 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644153 4744 feature_gate.go:330] unrecognized feature gate: Example Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644157 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644161 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644164 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644170 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644177 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644181 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644186 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644190 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644194 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644199 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644204 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644208 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644212 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644216 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644221 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644225 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644229 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644232 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644235 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644239 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644243 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644246 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644250 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644254 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644257 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644261 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644264 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644268 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644272 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644275 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.644279 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.645221 4744 flags.go:64] FLAG: --address="0.0.0.0" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.645239 4744 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.645250 4744 flags.go:64] FLAG: --anonymous-auth="true" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.646672 4744 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647086 4744 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647093 4744 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647101 4744 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647110 4744 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647118 4744 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647126 4744 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647135 4744 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647282 4744 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647310 4744 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647325 4744 flags.go:64] FLAG: --cgroup-root="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647336 4744 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647349 4744 flags.go:64] FLAG: --client-ca-file="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647361 4744 flags.go:64] FLAG: --cloud-config="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647371 4744 flags.go:64] FLAG: --cloud-provider="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647381 4744 flags.go:64] FLAG: --cluster-dns="[]" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647403 4744 flags.go:64] FLAG: --cluster-domain="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647423 4744 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647441 4744 flags.go:64] FLAG: --config-dir="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647840 4744 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647880 4744 flags.go:64] FLAG: --container-log-max-files="5" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647899 4744 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647910 4744 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647921 4744 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647932 4744 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647942 4744 flags.go:64] FLAG: --contention-profiling="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647952 4744 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647963 4744 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647976 4744 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.647986 4744 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648036 4744 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648048 4744 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648057 4744 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648067 4744 flags.go:64] FLAG: --enable-load-reader="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648078 4744 flags.go:64] FLAG: --enable-server="true" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648088 4744 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648106 4744 flags.go:64] FLAG: --event-burst="100" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648121 4744 flags.go:64] FLAG: --event-qps="50" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648134 4744 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648149 4744 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648162 4744 flags.go:64] FLAG: --eviction-hard="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648180 4744 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648195 4744 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648206 4744 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648217 4744 flags.go:64] FLAG: --eviction-soft="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648228 4744 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648238 4744 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648248 4744 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648258 4744 flags.go:64] FLAG: --experimental-mounter-path="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648267 4744 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648277 4744 flags.go:64] FLAG: --fail-swap-on="true" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648286 4744 flags.go:64] FLAG: --feature-gates="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648323 4744 flags.go:64] FLAG: --file-check-frequency="20s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648332 4744 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648343 4744 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648353 4744 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648365 4744 flags.go:64] FLAG: --healthz-port="10248" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648375 4744 flags.go:64] FLAG: --help="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648385 4744 flags.go:64] FLAG: --hostname-override="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648395 4744 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648404 4744 flags.go:64] FLAG: --http-check-frequency="20s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648414 4744 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648423 4744 flags.go:64] FLAG: --image-credential-provider-config="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648433 4744 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648442 4744 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648451 4744 flags.go:64] FLAG: --image-service-endpoint="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648461 4744 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648470 4744 flags.go:64] FLAG: --kube-api-burst="100" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648480 4744 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648522 4744 flags.go:64] FLAG: --kube-api-qps="50" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648534 4744 flags.go:64] FLAG: --kube-reserved="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648543 4744 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648553 4744 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648563 4744 flags.go:64] FLAG: --kubelet-cgroups="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648573 4744 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648582 4744 flags.go:64] FLAG: --lock-file="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648591 4744 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648601 4744 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648610 4744 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648626 4744 flags.go:64] FLAG: --log-json-split-stream="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648635 4744 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648645 4744 flags.go:64] FLAG: --log-text-split-stream="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648654 4744 flags.go:64] FLAG: --logging-format="text" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648664 4744 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648674 4744 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648683 4744 flags.go:64] FLAG: --manifest-url="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648696 4744 flags.go:64] FLAG: --manifest-url-header="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648713 4744 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648723 4744 flags.go:64] FLAG: --max-open-files="1000000" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648735 4744 flags.go:64] FLAG: --max-pods="110" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648744 4744 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648753 4744 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648764 4744 flags.go:64] FLAG: --memory-manager-policy="None" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648773 4744 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648783 4744 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648794 4744 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648803 4744 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648834 4744 flags.go:64] FLAG: --node-status-max-images="50" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648843 4744 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648853 4744 flags.go:64] FLAG: --oom-score-adj="-999" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648862 4744 flags.go:64] FLAG: --pod-cidr="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648871 4744 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648884 4744 flags.go:64] FLAG: --pod-manifest-path="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648893 4744 flags.go:64] FLAG: --pod-max-pids="-1" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648903 4744 flags.go:64] FLAG: --pods-per-core="0" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648912 4744 flags.go:64] FLAG: --port="10250" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648921 4744 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648930 4744 flags.go:64] FLAG: --provider-id="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648940 4744 flags.go:64] FLAG: --qos-reserved="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648950 4744 flags.go:64] FLAG: --read-only-port="10255" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648959 4744 flags.go:64] FLAG: --register-node="true" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648968 4744 flags.go:64] FLAG: --register-schedulable="true" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648977 4744 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.648996 4744 flags.go:64] FLAG: --registry-burst="10" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649005 4744 flags.go:64] FLAG: --registry-qps="5" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649014 4744 flags.go:64] FLAG: --reserved-cpus="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649023 4744 flags.go:64] FLAG: --reserved-memory="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649034 4744 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649044 4744 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649054 4744 flags.go:64] FLAG: --rotate-certificates="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649063 4744 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649072 4744 flags.go:64] FLAG: --runonce="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649081 4744 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649091 4744 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649100 4744 flags.go:64] FLAG: --seccomp-default="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649110 4744 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649121 4744 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649133 4744 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649146 4744 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649159 4744 flags.go:64] FLAG: --storage-driver-password="root" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649170 4744 flags.go:64] FLAG: --storage-driver-secure="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649182 4744 flags.go:64] FLAG: --storage-driver-table="stats" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649193 4744 flags.go:64] FLAG: --storage-driver-user="root" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649202 4744 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649212 4744 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649222 4744 flags.go:64] FLAG: --system-cgroups="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649231 4744 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649247 4744 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649255 4744 flags.go:64] FLAG: --tls-cert-file="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649264 4744 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649280 4744 flags.go:64] FLAG: --tls-min-version="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649290 4744 flags.go:64] FLAG: --tls-private-key-file="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649299 4744 flags.go:64] FLAG: --topology-manager-policy="none" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649308 4744 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649317 4744 flags.go:64] FLAG: --topology-manager-scope="container" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649327 4744 flags.go:64] FLAG: --v="2" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649340 4744 flags.go:64] FLAG: --version="false" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649352 4744 flags.go:64] FLAG: --vmodule="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649363 4744 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.649373 4744 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649660 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649672 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649681 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649690 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649701 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649709 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649717 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649725 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649737 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649747 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649757 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649767 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649776 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649785 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649793 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649801 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649812 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649822 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649830 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649839 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649848 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649857 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649866 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649875 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649884 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649894 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649904 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649913 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649921 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649929 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649938 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649947 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649956 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649964 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649972 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649981 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649988 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.649996 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650004 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650012 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650021 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650031 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650041 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650050 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650058 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650066 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650074 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650081 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650091 4744 feature_gate.go:330] unrecognized feature gate: Example Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650098 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650106 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650113 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650121 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650129 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650138 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650146 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650154 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650162 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650170 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650178 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650185 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650193 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650200 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650208 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650216 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650225 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650233 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650241 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650248 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650256 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.650264 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.650280 4744 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.662357 4744 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.662428 4744 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662615 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662639 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662649 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662659 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662667 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662676 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662685 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662694 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662707 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662722 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662734 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662744 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662755 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662764 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662773 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662783 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662794 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662802 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662811 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662819 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662827 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662835 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662844 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662852 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662860 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662869 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662877 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662885 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662894 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662902 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662910 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662919 4744 feature_gate.go:330] unrecognized feature gate: Example Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662927 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662936 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662947 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662956 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662964 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662975 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662986 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.662994 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663002 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663011 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663019 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663027 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663036 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663044 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663052 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663060 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663069 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663078 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663086 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663096 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663105 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663113 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663122 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663130 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663138 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663147 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663155 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663164 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663172 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663180 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663189 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663200 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663211 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663220 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663229 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663238 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663246 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663255 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.663267 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.663283 4744 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664601 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664625 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664637 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664647 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664656 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664666 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664674 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664684 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664693 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664701 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664710 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664719 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664728 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664737 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664749 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664760 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664770 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664781 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664793 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664804 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664813 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664824 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664833 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664842 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664852 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664862 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664872 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664882 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664891 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664900 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664909 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664918 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664927 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664936 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664952 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664965 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.664989 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665007 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665020 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665031 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665043 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665054 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665065 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665075 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665085 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665097 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665107 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665117 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665127 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665138 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665148 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665158 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665168 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665218 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665232 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665245 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665256 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665269 4744 feature_gate.go:330] unrecognized feature gate: Example Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665279 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665288 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665296 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665307 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665315 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665323 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665332 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665340 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665348 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665357 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665365 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665374 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.665389 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.665403 4744 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.665752 4744 server.go:940] "Client rotation is on, will bootstrap in background" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.672484 4744 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.672735 4744 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.677023 4744 server.go:997] "Starting client certificate rotation" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.677068 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.677323 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-16 13:16:33.726147865 +0000 UTC Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.677527 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1052h51m59.048662311s for next certificate rotation Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.711397 4744 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.715078 4744 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.732806 4744 log.go:25] "Validated CRI v1 runtime API" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.773508 4744 log.go:25] "Validated CRI v1 image API" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.775637 4744 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.788736 4744 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-03-16-19-36-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.788780 4744 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.807996 4744 manager.go:217] Machine: {Timestamp:2025-10-03 16:24:34.804669889 +0000 UTC m=+1.084545805 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:1aea8ebf-25a8-4f05-acb0-bcaaf1af2496 BootID:5c82bbbe-55e2-4b37-ab80-2af91bc4605d Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:92:5d:94 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:92:5d:94 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:85:0f:94 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:b6:83:27 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:55:1a:cf Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:7b:bf:ce Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ee:02:40:a2:28:f7 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:d6:90:91:68:68:70 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.808351 4744 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.808725 4744 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.811727 4744 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.812158 4744 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.812247 4744 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.812647 4744 topology_manager.go:138] "Creating topology manager with none policy" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.812669 4744 container_manager_linux.go:303] "Creating device plugin manager" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.813353 4744 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.813412 4744 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.813831 4744 state_mem.go:36] "Initialized new in-memory state store" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.814405 4744 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.818041 4744 kubelet.go:418] "Attempting to sync node with API server" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.818078 4744 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.818099 4744 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.818116 4744 kubelet.go:324] "Adding apiserver pod source" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.818134 4744 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.824971 4744 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.825323 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.132:6443: connect: connection refused Oct 03 16:24:34 crc kubenswrapper[4744]: E1003 16:24:34.825489 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.132:6443: connect: connection refused" logger="UnhandledError" Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.825472 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.132:6443: connect: connection refused Oct 03 16:24:34 crc kubenswrapper[4744]: E1003 16:24:34.825684 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.132:6443: connect: connection refused" logger="UnhandledError" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.825783 4744 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.828143 4744 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.829691 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.829717 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.829726 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.829734 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.829745 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.829752 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.829760 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.829772 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.829783 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.829790 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.829815 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.829822 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.830768 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.831674 4744 server.go:1280] "Started kubelet" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.832689 4744 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.832663 4744 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.833143 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.132:6443: connect: connection refused Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.833863 4744 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 03 16:24:34 crc systemd[1]: Started Kubernetes Kubelet. Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.834395 4744 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.834427 4744 server.go:460] "Adding debug handlers to kubelet server" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.834448 4744 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 19:15:54.524536782 +0000 UTC Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.834515 4744 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2258h51m19.690025025s for next certificate rotation Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.834429 4744 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.835075 4744 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.835127 4744 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.835313 4744 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 03 16:24:34 crc kubenswrapper[4744]: E1003 16:24:34.835380 4744 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 03 16:24:34 crc kubenswrapper[4744]: E1003 16:24:34.838346 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.132:6443: connect: connection refused" interval="200ms" Oct 03 16:24:34 crc kubenswrapper[4744]: E1003 16:24:34.840204 4744 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.132:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b07d1468ee0b2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-03 16:24:34.831638706 +0000 UTC m=+1.111514602,LastTimestamp:2025-10-03 16:24:34.831638706 +0000 UTC m=+1.111514602,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.843032 4744 factory.go:55] Registering systemd factory Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.843277 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.132:6443: connect: connection refused Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.843414 4744 factory.go:221] Registration of the systemd container factory successfully Oct 03 16:24:34 crc kubenswrapper[4744]: E1003 16:24:34.843563 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.132:6443: connect: connection refused" logger="UnhandledError" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.844396 4744 factory.go:153] Registering CRI-O factory Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.845540 4744 factory.go:221] Registration of the crio container factory successfully Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.845718 4744 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.845807 4744 factory.go:103] Registering Raw factory Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.845844 4744 manager.go:1196] Started watching for new ooms in manager Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.846797 4744 manager.go:319] Starting recovery of all containers Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851294 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851407 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851420 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851434 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851445 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851459 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851474 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851485 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851520 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851530 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851542 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851554 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851564 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851582 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851596 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851612 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851626 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851636 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851648 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851664 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851675 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851686 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851698 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851712 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851725 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851739 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851755 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851797 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851812 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851824 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851859 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851873 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851886 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851895 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851904 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851913 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851923 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851931 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851940 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851951 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851960 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851970 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851981 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.851994 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852007 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852019 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852032 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852042 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852052 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852062 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852072 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852081 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852096 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852108 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852118 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852127 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852137 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852149 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852159 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852169 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852182 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852192 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852203 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852212 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852223 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852233 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852243 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852252 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852262 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852273 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852282 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852292 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852330 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852342 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852353 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852363 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852373 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852383 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852392 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852402 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852412 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852421 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852432 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852443 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852454 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852465 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852478 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852503 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852513 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852522 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852532 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852540 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852553 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852567 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852579 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852593 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852605 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852617 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852628 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852640 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852652 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852665 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852675 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852689 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852704 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852716 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852727 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852739 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852749 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852759 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852770 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852782 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852795 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852806 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852818 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852829 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852842 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852857 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852870 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852881 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852891 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852900 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852911 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852925 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852940 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852950 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852960 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852970 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852982 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.852991 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853002 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853012 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853024 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853037 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853048 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853061 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853075 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853087 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853101 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853114 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853126 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853140 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853152 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853161 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853175 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853187 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.853203 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855443 4744 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855482 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855520 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855535 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855550 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855567 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855580 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855593 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855606 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855618 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855629 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855641 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855656 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855672 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855685 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855698 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855711 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855725 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855738 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855750 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855761 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855775 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855788 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855800 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855811 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855823 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855838 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855850 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855864 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855880 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855893 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855906 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855918 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855932 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855945 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.855973 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856035 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856051 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856066 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856078 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856093 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856105 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856117 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856131 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856145 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856164 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856176 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856188 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856200 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856212 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856224 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856237 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856253 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856270 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856285 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856299 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856313 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856326 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856339 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856351 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856364 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856379 4744 reconstruct.go:97] "Volume reconstruction finished" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.856389 4744 reconciler.go:26] "Reconciler: start to sync state" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.869276 4744 manager.go:324] Recovery completed Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.887192 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.888134 4744 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.890297 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.890349 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.890361 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.890582 4744 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.890630 4744 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.890654 4744 kubelet.go:2335] "Starting kubelet main sync loop" Oct 03 16:24:34 crc kubenswrapper[4744]: E1003 16:24:34.890697 4744 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.891255 4744 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.891276 4744 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.891302 4744 state_mem.go:36] "Initialized new in-memory state store" Oct 03 16:24:34 crc kubenswrapper[4744]: W1003 16:24:34.892795 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.132:6443: connect: connection refused Oct 03 16:24:34 crc kubenswrapper[4744]: E1003 16:24:34.892861 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.132:6443: connect: connection refused" logger="UnhandledError" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.909213 4744 policy_none.go:49] "None policy: Start" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.910521 4744 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.910557 4744 state_mem.go:35] "Initializing new in-memory state store" Oct 03 16:24:34 crc kubenswrapper[4744]: E1003 16:24:34.935678 4744 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.971035 4744 manager.go:334] "Starting Device Plugin manager" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.971127 4744 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.971141 4744 server.go:79] "Starting device plugin registration server" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.971580 4744 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.971594 4744 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.971805 4744 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.971873 4744 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.971881 4744 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 03 16:24:34 crc kubenswrapper[4744]: E1003 16:24:34.982334 4744 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.991693 4744 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.991805 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.993043 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.993085 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.993101 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.993270 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.993760 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.993864 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.998880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.998950 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.998961 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.999192 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.999239 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.999258 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.999445 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.999738 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 16:24:34 crc kubenswrapper[4744]: I1003 16:24:34.999855 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.000875 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.001012 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.001033 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.001264 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.001435 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.001531 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.001700 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.001755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.001774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.002660 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.002702 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.002716 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.002859 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.003025 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.003074 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.003209 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.003237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.003246 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.003904 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.003934 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.003944 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.004133 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.004172 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.004192 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.004429 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.004465 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.005557 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.005601 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.005621 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:35 crc kubenswrapper[4744]: E1003 16:24:35.039419 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.132:6443: connect: connection refused" interval="400ms" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.059044 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.059104 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.059135 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.059158 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.059235 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.059309 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.059380 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.059435 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.059469 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.059508 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.059533 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.059553 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.059573 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.059642 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.059677 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.071790 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.073047 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.073095 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.073107 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.073150 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 16:24:35 crc kubenswrapper[4744]: E1003 16:24:35.073836 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.132:6443: connect: connection refused" node="crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161212 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161268 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161290 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161315 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161335 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161353 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161371 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161391 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161410 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161421 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161511 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161554 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161453 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161469 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161568 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161423 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161536 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161647 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161628 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161679 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161745 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161725 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161870 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161928 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161941 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161965 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.161984 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.162027 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.162098 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.162120 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.274582 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.276348 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.276386 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.276398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.276425 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 16:24:35 crc kubenswrapper[4744]: E1003 16:24:35.277062 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.132:6443: connect: connection refused" node="crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.341278 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.346094 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.367828 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.374292 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.378769 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:35 crc kubenswrapper[4744]: W1003 16:24:35.393360 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-fee0b49ccecae57434ed21ecb25a7eb58d862c8aad41eb73e0e2812ebd660c8f WatchSource:0}: Error finding container fee0b49ccecae57434ed21ecb25a7eb58d862c8aad41eb73e0e2812ebd660c8f: Status 404 returned error can't find the container with id fee0b49ccecae57434ed21ecb25a7eb58d862c8aad41eb73e0e2812ebd660c8f Oct 03 16:24:35 crc kubenswrapper[4744]: W1003 16:24:35.394845 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-30ef8bf9a1da7eb93245cc36e3e91b53b1ef77e8b6de42ea5956ad350ce70e38 WatchSource:0}: Error finding container 30ef8bf9a1da7eb93245cc36e3e91b53b1ef77e8b6de42ea5956ad350ce70e38: Status 404 returned error can't find the container with id 30ef8bf9a1da7eb93245cc36e3e91b53b1ef77e8b6de42ea5956ad350ce70e38 Oct 03 16:24:35 crc kubenswrapper[4744]: W1003 16:24:35.403676 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-0b8c43047b4f5e11d682296202589f49e2937118272866a5967472503831fdbd WatchSource:0}: Error finding container 0b8c43047b4f5e11d682296202589f49e2937118272866a5967472503831fdbd: Status 404 returned error can't find the container with id 0b8c43047b4f5e11d682296202589f49e2937118272866a5967472503831fdbd Oct 03 16:24:35 crc kubenswrapper[4744]: W1003 16:24:35.411266 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-cd514b92b8742fdcaa3838e5ea52a10c037d93b90826f9a0080ab68bf5ab7faf WatchSource:0}: Error finding container cd514b92b8742fdcaa3838e5ea52a10c037d93b90826f9a0080ab68bf5ab7faf: Status 404 returned error can't find the container with id cd514b92b8742fdcaa3838e5ea52a10c037d93b90826f9a0080ab68bf5ab7faf Oct 03 16:24:35 crc kubenswrapper[4744]: W1003 16:24:35.413535 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-6811faa7db19d6d1b8dcb3d81ded51d9d5ea8b5e11072fc82a88acdf05c7d313 WatchSource:0}: Error finding container 6811faa7db19d6d1b8dcb3d81ded51d9d5ea8b5e11072fc82a88acdf05c7d313: Status 404 returned error can't find the container with id 6811faa7db19d6d1b8dcb3d81ded51d9d5ea8b5e11072fc82a88acdf05c7d313 Oct 03 16:24:35 crc kubenswrapper[4744]: E1003 16:24:35.440753 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.132:6443: connect: connection refused" interval="800ms" Oct 03 16:24:35 crc kubenswrapper[4744]: W1003 16:24:35.649175 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.132:6443: connect: connection refused Oct 03 16:24:35 crc kubenswrapper[4744]: E1003 16:24:35.649362 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.132:6443: connect: connection refused" logger="UnhandledError" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.677702 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.679812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.679865 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.679878 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.679913 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 16:24:35 crc kubenswrapper[4744]: E1003 16:24:35.680629 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.132:6443: connect: connection refused" node="crc" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.834110 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.132:6443: connect: connection refused Oct 03 16:24:35 crc kubenswrapper[4744]: W1003 16:24:35.862278 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.132:6443: connect: connection refused Oct 03 16:24:35 crc kubenswrapper[4744]: E1003 16:24:35.862394 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.132:6443: connect: connection refused" logger="UnhandledError" Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.899454 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6811faa7db19d6d1b8dcb3d81ded51d9d5ea8b5e11072fc82a88acdf05c7d313"} Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.900671 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cd514b92b8742fdcaa3838e5ea52a10c037d93b90826f9a0080ab68bf5ab7faf"} Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.901980 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0b8c43047b4f5e11d682296202589f49e2937118272866a5967472503831fdbd"} Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.903871 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"30ef8bf9a1da7eb93245cc36e3e91b53b1ef77e8b6de42ea5956ad350ce70e38"} Oct 03 16:24:35 crc kubenswrapper[4744]: I1003 16:24:35.905891 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fee0b49ccecae57434ed21ecb25a7eb58d862c8aad41eb73e0e2812ebd660c8f"} Oct 03 16:24:36 crc kubenswrapper[4744]: W1003 16:24:36.020295 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.132:6443: connect: connection refused Oct 03 16:24:36 crc kubenswrapper[4744]: E1003 16:24:36.020796 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.132:6443: connect: connection refused" logger="UnhandledError" Oct 03 16:24:36 crc kubenswrapper[4744]: E1003 16:24:36.242161 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.132:6443: connect: connection refused" interval="1.6s" Oct 03 16:24:36 crc kubenswrapper[4744]: W1003 16:24:36.354850 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.132:6443: connect: connection refused Oct 03 16:24:36 crc kubenswrapper[4744]: E1003 16:24:36.355036 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.132:6443: connect: connection refused" logger="UnhandledError" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.481113 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.482636 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.482692 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.482707 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.482738 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 16:24:36 crc kubenswrapper[4744]: E1003 16:24:36.483125 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.132:6443: connect: connection refused" node="crc" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.834081 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.132:6443: connect: connection refused Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.912553 4744 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d" exitCode=0 Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.912647 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d"} Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.912719 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.913781 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.913825 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.913839 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.914777 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b" exitCode=0 Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.914848 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b"} Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.914889 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.915744 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.915764 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.915775 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.917310 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.917815 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5"} Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.917852 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651"} Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.917865 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2"} Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.917911 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce"} Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.917929 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.918316 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.918352 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.918361 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.918776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.918803 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.918814 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.919727 4744 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="2a3b07aa960b742c202322c07932081b0fcc88f406dd1adc5c33562376d7b9ae" exitCode=0 Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.919789 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"2a3b07aa960b742c202322c07932081b0fcc88f406dd1adc5c33562376d7b9ae"} Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.919797 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.920529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.920559 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.920571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.921643 4744 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458" exitCode=0 Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.921695 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458"} Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.921767 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.922427 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.922470 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:36 crc kubenswrapper[4744]: I1003 16:24:36.922484 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.834702 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.132:6443: connect: connection refused Oct 03 16:24:37 crc kubenswrapper[4744]: E1003 16:24:37.844125 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.132:6443: connect: connection refused" interval="3.2s" Oct 03 16:24:37 crc kubenswrapper[4744]: E1003 16:24:37.926023 4744 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.132:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b07d1468ee0b2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-03 16:24:34.831638706 +0000 UTC m=+1.111514602,LastTimestamp:2025-10-03 16:24:34.831638706 +0000 UTC m=+1.111514602,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.928385 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8"} Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.928431 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509"} Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.928441 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64"} Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.928451 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19"} Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.928461 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030"} Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.928596 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.929799 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.929828 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.929839 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.932751 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.932875 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a475eeb458ca840cc4a5d5c82b30a55bd6f245681131113878dc984915a9dc31"} Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.934333 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.934627 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.934739 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.935993 4744 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0" exitCode=0 Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.936096 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0"} Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.936182 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.937072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.937255 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.937321 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.941144 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.941619 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.941967 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"52cbfb9a1c58ce525029b4226599a8e258453a8c5837643a6fd1efe2441641a8"} Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.942004 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7e6b1c34f972682f8afe389a96e2b35b848e4ceb02cb9379fa91c764f2cf11e4"} Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.942019 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"011b2e41b97e95ff1a11ff71d90fc00813ab18eac48b6cc7a987039494649a15"} Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.942471 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.942511 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.942525 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.942558 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.942600 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:37 crc kubenswrapper[4744]: I1003 16:24:37.942616 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.084315 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.086656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.086705 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.086719 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.086744 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 16:24:38 crc kubenswrapper[4744]: E1003 16:24:38.087376 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.132:6443: connect: connection refused" node="crc" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.135691 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.945475 4744 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319" exitCode=0 Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.945634 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.945660 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319"} Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.945680 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.945694 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.945699 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.945842 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.945845 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.946575 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.946895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.946937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.946952 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.947089 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.947096 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.947114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.947120 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.947123 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.947141 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.947149 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.947162 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.947132 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.947945 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.948044 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:38 crc kubenswrapper[4744]: I1003 16:24:38.948130 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:39 crc kubenswrapper[4744]: I1003 16:24:39.156997 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:39 crc kubenswrapper[4744]: I1003 16:24:39.573488 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:39 crc kubenswrapper[4744]: I1003 16:24:39.952114 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0"} Oct 03 16:24:39 crc kubenswrapper[4744]: I1003 16:24:39.952172 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134"} Oct 03 16:24:39 crc kubenswrapper[4744]: I1003 16:24:39.952207 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:39 crc kubenswrapper[4744]: I1003 16:24:39.952979 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:39 crc kubenswrapper[4744]: I1003 16:24:39.953004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:39 crc kubenswrapper[4744]: I1003 16:24:39.953014 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.194316 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.194543 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.195719 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.195871 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.195934 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.957052 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8"} Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.957099 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d"} Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.957112 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2"} Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.957128 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.957128 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.957954 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.957968 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.957986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.957986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.958010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:40 crc kubenswrapper[4744]: I1003 16:24:40.957998 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:41 crc kubenswrapper[4744]: I1003 16:24:41.136093 4744 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 03 16:24:41 crc kubenswrapper[4744]: I1003 16:24:41.136182 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 03 16:24:41 crc kubenswrapper[4744]: I1003 16:24:41.288083 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:41 crc kubenswrapper[4744]: I1003 16:24:41.289529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:41 crc kubenswrapper[4744]: I1003 16:24:41.289591 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:41 crc kubenswrapper[4744]: I1003 16:24:41.289605 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:41 crc kubenswrapper[4744]: I1003 16:24:41.289634 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 16:24:41 crc kubenswrapper[4744]: I1003 16:24:41.960304 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:41 crc kubenswrapper[4744]: I1003 16:24:41.961619 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:41 crc kubenswrapper[4744]: I1003 16:24:41.961700 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:41 crc kubenswrapper[4744]: I1003 16:24:41.961720 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:42 crc kubenswrapper[4744]: I1003 16:24:42.124942 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:42 crc kubenswrapper[4744]: I1003 16:24:42.125304 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:42 crc kubenswrapper[4744]: I1003 16:24:42.127079 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:42 crc kubenswrapper[4744]: I1003 16:24:42.127113 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:42 crc kubenswrapper[4744]: I1003 16:24:42.127126 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:43 crc kubenswrapper[4744]: I1003 16:24:43.849088 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 16:24:43 crc kubenswrapper[4744]: I1003 16:24:43.849397 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:43 crc kubenswrapper[4744]: I1003 16:24:43.851062 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:43 crc kubenswrapper[4744]: I1003 16:24:43.851116 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:43 crc kubenswrapper[4744]: I1003 16:24:43.851138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:44 crc kubenswrapper[4744]: I1003 16:24:44.105535 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:44 crc kubenswrapper[4744]: I1003 16:24:44.105799 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:44 crc kubenswrapper[4744]: I1003 16:24:44.107454 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:44 crc kubenswrapper[4744]: I1003 16:24:44.107511 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:44 crc kubenswrapper[4744]: I1003 16:24:44.107523 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:44 crc kubenswrapper[4744]: I1003 16:24:44.407989 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 03 16:24:44 crc kubenswrapper[4744]: I1003 16:24:44.408312 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:44 crc kubenswrapper[4744]: I1003 16:24:44.410174 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:44 crc kubenswrapper[4744]: I1003 16:24:44.410216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:44 crc kubenswrapper[4744]: I1003 16:24:44.410234 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:44 crc kubenswrapper[4744]: E1003 16:24:44.982476 4744 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 03 16:24:45 crc kubenswrapper[4744]: I1003 16:24:45.901242 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:45 crc kubenswrapper[4744]: I1003 16:24:45.901426 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:45 crc kubenswrapper[4744]: I1003 16:24:45.902516 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:45 crc kubenswrapper[4744]: I1003 16:24:45.902548 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:45 crc kubenswrapper[4744]: I1003 16:24:45.902560 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:45 crc kubenswrapper[4744]: I1003 16:24:45.906283 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:45 crc kubenswrapper[4744]: I1003 16:24:45.971306 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:45 crc kubenswrapper[4744]: I1003 16:24:45.972217 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:45 crc kubenswrapper[4744]: I1003 16:24:45.972257 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:45 crc kubenswrapper[4744]: I1003 16:24:45.972266 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:47 crc kubenswrapper[4744]: I1003 16:24:47.273889 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 03 16:24:47 crc kubenswrapper[4744]: I1003 16:24:47.274095 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:47 crc kubenswrapper[4744]: I1003 16:24:47.275461 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:47 crc kubenswrapper[4744]: I1003 16:24:47.275527 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:47 crc kubenswrapper[4744]: I1003 16:24:47.275539 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:48 crc kubenswrapper[4744]: W1003 16:24:48.478776 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 03 16:24:48 crc kubenswrapper[4744]: I1003 16:24:48.478934 4744 trace.go:236] Trace[863257626]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Oct-2025 16:24:38.476) (total time: 10002ms): Oct 03 16:24:48 crc kubenswrapper[4744]: Trace[863257626]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (16:24:48.478) Oct 03 16:24:48 crc kubenswrapper[4744]: Trace[863257626]: [10.002012911s] [10.002012911s] END Oct 03 16:24:48 crc kubenswrapper[4744]: E1003 16:24:48.478974 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 03 16:24:48 crc kubenswrapper[4744]: W1003 16:24:48.603020 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 03 16:24:48 crc kubenswrapper[4744]: I1003 16:24:48.603160 4744 trace.go:236] Trace[1246056059]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Oct-2025 16:24:38.601) (total time: 10002ms): Oct 03 16:24:48 crc kubenswrapper[4744]: Trace[1246056059]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (16:24:48.603) Oct 03 16:24:48 crc kubenswrapper[4744]: Trace[1246056059]: [10.002104834s] [10.002104834s] END Oct 03 16:24:48 crc kubenswrapper[4744]: E1003 16:24:48.603200 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 03 16:24:48 crc kubenswrapper[4744]: W1003 16:24:48.786137 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 03 16:24:48 crc kubenswrapper[4744]: I1003 16:24:48.786289 4744 trace.go:236] Trace[1159702250]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Oct-2025 16:24:38.784) (total time: 10001ms): Oct 03 16:24:48 crc kubenswrapper[4744]: Trace[1159702250]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (16:24:48.786) Oct 03 16:24:48 crc kubenswrapper[4744]: Trace[1159702250]: [10.001596921s] [10.001596921s] END Oct 03 16:24:48 crc kubenswrapper[4744]: E1003 16:24:48.786328 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 03 16:24:48 crc kubenswrapper[4744]: W1003 16:24:48.804956 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 03 16:24:48 crc kubenswrapper[4744]: I1003 16:24:48.805086 4744 trace.go:236] Trace[1520174139]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Oct-2025 16:24:38.803) (total time: 10001ms): Oct 03 16:24:48 crc kubenswrapper[4744]: Trace[1520174139]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (16:24:48.804) Oct 03 16:24:48 crc kubenswrapper[4744]: Trace[1520174139]: [10.001662662s] [10.001662662s] END Oct 03 16:24:48 crc kubenswrapper[4744]: E1003 16:24:48.805115 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 03 16:24:48 crc kubenswrapper[4744]: I1003 16:24:48.835140 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 03 16:24:48 crc kubenswrapper[4744]: I1003 16:24:48.986094 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 03 16:24:48 crc kubenswrapper[4744]: I1003 16:24:48.989254 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8" exitCode=255 Oct 03 16:24:48 crc kubenswrapper[4744]: I1003 16:24:48.989370 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8"} Oct 03 16:24:48 crc kubenswrapper[4744]: I1003 16:24:48.990469 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:48 crc kubenswrapper[4744]: I1003 16:24:48.991891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:48 crc kubenswrapper[4744]: I1003 16:24:48.991949 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:48 crc kubenswrapper[4744]: I1003 16:24:48.991963 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:48 crc kubenswrapper[4744]: I1003 16:24:48.992762 4744 scope.go:117] "RemoveContainer" containerID="8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8" Oct 03 16:24:49 crc kubenswrapper[4744]: I1003 16:24:49.714577 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 03 16:24:49 crc kubenswrapper[4744]: I1003 16:24:49.714669 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 03 16:24:49 crc kubenswrapper[4744]: I1003 16:24:49.720715 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 03 16:24:49 crc kubenswrapper[4744]: I1003 16:24:49.720823 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 03 16:24:49 crc kubenswrapper[4744]: I1003 16:24:49.998992 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 03 16:24:50 crc kubenswrapper[4744]: I1003 16:24:50.000465 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba"} Oct 03 16:24:50 crc kubenswrapper[4744]: I1003 16:24:50.000671 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:50 crc kubenswrapper[4744]: I1003 16:24:50.001627 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:50 crc kubenswrapper[4744]: I1003 16:24:50.001695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:50 crc kubenswrapper[4744]: I1003 16:24:50.001714 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:50 crc kubenswrapper[4744]: I1003 16:24:50.206923 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:50 crc kubenswrapper[4744]: I1003 16:24:50.207144 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:50 crc kubenswrapper[4744]: I1003 16:24:50.208954 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:50 crc kubenswrapper[4744]: I1003 16:24:50.209012 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:50 crc kubenswrapper[4744]: I1003 16:24:50.209021 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:51 crc kubenswrapper[4744]: I1003 16:24:51.129891 4744 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 03 16:24:51 crc kubenswrapper[4744]: I1003 16:24:51.130688 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 03 16:24:52 crc kubenswrapper[4744]: I1003 16:24:52.133469 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:52 crc kubenswrapper[4744]: I1003 16:24:52.133688 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:24:52 crc kubenswrapper[4744]: I1003 16:24:52.133892 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:52 crc kubenswrapper[4744]: I1003 16:24:52.135633 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:24:52 crc kubenswrapper[4744]: I1003 16:24:52.135709 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:24:52 crc kubenswrapper[4744]: I1003 16:24:52.135737 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:24:52 crc kubenswrapper[4744]: I1003 16:24:52.141295 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:52 crc kubenswrapper[4744]: I1003 16:24:52.793597 4744 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 03 16:24:52 crc kubenswrapper[4744]: I1003 16:24:52.923322 4744 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.313812 4744 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.799686 4744 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.827021 4744 apiserver.go:52] "Watching apiserver" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.831212 4744 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.831663 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.832082 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.832111 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.832182 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:24:53 crc kubenswrapper[4744]: E1003 16:24:53.832179 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.832758 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 16:24:53 crc kubenswrapper[4744]: E1003 16:24:53.832801 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.832935 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.833361 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:24:53 crc kubenswrapper[4744]: E1003 16:24:53.833484 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.835638 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.836774 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.836921 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.838622 4744 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.839839 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.840317 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.841990 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.842151 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.842152 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.842309 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.868189 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.879294 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.895328 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.913414 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.929327 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.941644 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.953487 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:53 crc kubenswrapper[4744]: I1003 16:24:53.966334 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.023580 4744 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.708306 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.711798 4744 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.712104 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.812436 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.812510 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.812543 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.812574 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.812697 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.812728 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.812754 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.812784 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.812811 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.812838 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.812865 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.812892 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.813700 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.813674 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.813744 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.813924 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.813987 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.814075 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.814280 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.814296 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.814318 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.814404 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.814477 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.814690 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.814739 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.814830 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.814992 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.815067 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.815175 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.815258 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.815327 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.815401 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.815478 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.815585 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.815659 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.815930 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.816084 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.816224 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.816301 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.816381 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.815296 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.815418 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.815618 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.815796 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.815806 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.816779 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.815999 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.816170 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.816828 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.816361 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.816741 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.816750 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.816463 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:24:55.316432058 +0000 UTC m=+21.596307944 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.816936 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.817146 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.817228 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.817315 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.817228 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.817375 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.817590 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.817396 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.817884 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.817831 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.817864 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.818049 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.818129 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.818213 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.818284 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.818348 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.818414 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.818828 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.818246 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.818452 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.818699 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.818781 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.818786 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.819068 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.819079 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.818912 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.819291 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.819383 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.819465 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.819575 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.819591 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.819731 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.819804 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.819920 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820031 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820108 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820252 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.819650 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.819669 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.819837 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820297 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820339 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820341 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820278 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820374 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820529 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820557 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820573 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820582 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820614 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820643 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820668 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820694 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820721 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820747 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820777 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820802 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820833 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820858 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820888 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820915 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820945 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820971 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820998 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.820999 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.821006 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.821028 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.821064 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.821091 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.821120 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.821147 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.821173 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.821230 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.822465 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.822681 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.822298 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.822741 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.822789 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.822822 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.822847 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.822871 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.822896 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.822919 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.822971 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.822991 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823016 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823039 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823059 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823088 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823112 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823135 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823157 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823209 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823231 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823255 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823274 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823295 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823353 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823631 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823666 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823691 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823712 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823734 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823728 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823753 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823779 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823801 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823820 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823859 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823863 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823880 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823902 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823894 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823921 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823943 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823964 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.823982 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824004 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824028 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824049 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824065 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824089 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824118 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824137 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824168 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824172 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824192 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824212 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824217 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824236 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824260 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824283 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824301 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824325 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824352 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824373 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824393 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824413 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824430 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824450 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824470 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824510 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824485 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824535 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824562 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824581 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824602 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824626 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824651 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824668 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824689 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824710 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824727 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824747 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824766 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824767 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824786 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824805 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824824 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824847 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824866 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824886 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824907 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824926 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824942 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824962 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.824983 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825002 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825025 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825031 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825067 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825084 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825104 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825123 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825144 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825161 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825181 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825201 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825218 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825237 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825257 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825273 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825292 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825300 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825310 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825362 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825402 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825429 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825457 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825487 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825537 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825563 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825588 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825613 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825659 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825686 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825715 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825740 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825769 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825797 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825823 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825845 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825885 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825892 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825907 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825924 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.825944 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826003 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826032 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826054 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826077 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826098 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826143 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826182 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826203 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826226 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826291 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826322 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826350 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826387 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826523 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826540 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826561 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826576 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826590 4744 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826610 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826624 4744 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826639 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826652 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826671 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826687 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826701 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826720 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826733 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826746 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826773 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826792 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826808 4744 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826819 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826832 4744 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826849 4744 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826867 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826882 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826898 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826911 4744 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826928 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826942 4744 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826956 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826978 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.826993 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827004 4744 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827014 4744 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827029 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827040 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827050 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827060 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827076 4744 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827087 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827098 4744 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827112 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827122 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827131 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827141 4744 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827152 4744 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827162 4744 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827171 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827181 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827218 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827230 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827241 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827252 4744 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827266 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827277 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827289 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827303 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827313 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827323 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827334 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827339 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827348 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827434 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827452 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827468 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827457 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827480 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827480 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827534 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827553 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.827566 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.828398 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.831656 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.831877 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.831972 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.832017 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.832158 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.832257 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:24:55.332223557 +0000 UTC m=+21.612099453 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.832277 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.832337 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.832552 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.832664 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.832814 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.832973 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.833080 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.833124 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.833260 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.833297 4744 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.833356 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:24:55.333314404 +0000 UTC m=+21.613190300 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.833273 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.833475 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.833737 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.834103 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.834230 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.834410 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.834434 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.837520 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.834874 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.831094 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.835645 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.836863 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.837119 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.837585 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.834150 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.837672 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.837823 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.838024 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.838340 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.838357 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.838838 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.839533 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.839605 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.839988 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.840043 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.840739 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.847093 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.848565 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.848778 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.849211 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.849222 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.849214 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.849298 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.849972 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.849975 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.850291 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.850744 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.851801 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.851940 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.852251 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.852871 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.852915 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.853116 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.853203 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.853386 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.853417 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.853472 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.853657 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.853693 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.853989 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.854031 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.854246 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.854317 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.853013 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.855196 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.855716 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.855732 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.855970 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.855997 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.856523 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.856555 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.856709 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.856906 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.857247 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.857260 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.857590 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.857812 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.857996 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.858251 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.858463 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.859208 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.859292 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.859726 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.859748 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.859755 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.859796 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.859810 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.859932 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 16:24:55.359902783 +0000 UTC m=+21.639778679 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.859969 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.860028 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.860225 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.860081 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.857477 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.860362 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.860382 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.860413 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:54 crc kubenswrapper[4744]: E1003 16:24:54.860491 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 16:24:55.360454896 +0000 UTC m=+21.640330792 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.860138 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.860635 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.860724 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.858885 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.860878 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.861023 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.861351 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.861650 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.861822 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.862108 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.862106 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.861957 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.862478 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.862574 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.862588 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.862743 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.863127 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.863197 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.863433 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.864162 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.864346 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.864587 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.864736 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.865076 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.865126 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.865149 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.865801 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.866244 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.867238 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.867283 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.867293 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.867346 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.867576 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.868423 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.875927 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.877444 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.877625 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.879442 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.881583 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.898723 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.898810 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.903360 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.904371 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.905379 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.905828 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.906448 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.908191 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.908736 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.909478 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.915481 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.916280 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.917288 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.917938 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.920490 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.921028 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.921177 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.922273 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.922960 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.924409 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.925870 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.927144 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.927841 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.928982 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929029 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929094 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929116 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929131 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929146 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929158 4744 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929167 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929176 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929185 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929194 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929205 4744 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929242 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929255 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929263 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929271 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929282 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929291 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929300 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929308 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929317 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929325 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929334 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929343 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929352 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929361 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929369 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929378 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929387 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929396 4744 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929407 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929417 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929427 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929473 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929510 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929556 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929570 4744 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929580 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929584 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929590 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929599 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929608 4744 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929618 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929627 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929635 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929644 4744 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929653 4744 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929662 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929671 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929679 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929689 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929697 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929706 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929715 4744 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929733 4744 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929742 4744 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929749 4744 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929759 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929768 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929778 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929787 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929796 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929804 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929814 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929824 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929832 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929841 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929849 4744 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929857 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929865 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929874 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929885 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929893 4744 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929901 4744 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929909 4744 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929918 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929926 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929935 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929944 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929952 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929962 4744 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929972 4744 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929981 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929991 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.929999 4744 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930008 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930016 4744 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930029 4744 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930037 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930046 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930054 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930063 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930071 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930079 4744 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930089 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930097 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930105 4744 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930113 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930121 4744 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930129 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930146 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930155 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930163 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930172 4744 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930180 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930188 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930196 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930204 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930212 4744 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930225 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930233 4744 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930241 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930249 4744 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930258 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930266 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930275 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930284 4744 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930293 4744 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930302 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930310 4744 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930320 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930328 4744 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930337 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930356 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930365 4744 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930374 4744 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930384 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930392 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930401 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930409 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930417 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930427 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930436 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930444 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930452 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.930461 4744 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.931555 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.931675 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.932041 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.933084 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.933478 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.934825 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.935683 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.936241 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.937204 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.937821 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.938876 4744 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.939084 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.941732 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.943068 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.943658 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.943650 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.945942 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.946814 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.947995 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.948779 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.952041 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.952834 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.954318 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.956389 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.957063 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.957118 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.957584 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.959473 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.960304 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.961827 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.963460 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.964276 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.964964 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.966637 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.967369 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.969001 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.969647 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-pmktb"] Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.970085 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-pmktb" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.972355 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.972638 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.973001 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.974262 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:54 crc kubenswrapper[4744]: I1003 16:24:54.986523 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.002559 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.017342 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.034605 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.044161 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.050173 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.055740 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.064161 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.064341 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.080287 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.098219 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.118563 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.132373 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blqh6\" (UniqueName: \"kubernetes.io/projected/134d85a0-99b6-4e3d-8960-5b168792a0ba-kube-api-access-blqh6\") pod \"node-resolver-pmktb\" (UID: \"134d85a0-99b6-4e3d-8960-5b168792a0ba\") " pod="openshift-dns/node-resolver-pmktb" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.132460 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/134d85a0-99b6-4e3d-8960-5b168792a0ba-hosts-file\") pod \"node-resolver-pmktb\" (UID: \"134d85a0-99b6-4e3d-8960-5b168792a0ba\") " pod="openshift-dns/node-resolver-pmktb" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.233941 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/134d85a0-99b6-4e3d-8960-5b168792a0ba-hosts-file\") pod \"node-resolver-pmktb\" (UID: \"134d85a0-99b6-4e3d-8960-5b168792a0ba\") " pod="openshift-dns/node-resolver-pmktb" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.233984 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blqh6\" (UniqueName: \"kubernetes.io/projected/134d85a0-99b6-4e3d-8960-5b168792a0ba-kube-api-access-blqh6\") pod \"node-resolver-pmktb\" (UID: \"134d85a0-99b6-4e3d-8960-5b168792a0ba\") " pod="openshift-dns/node-resolver-pmktb" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.234326 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/134d85a0-99b6-4e3d-8960-5b168792a0ba-hosts-file\") pod \"node-resolver-pmktb\" (UID: \"134d85a0-99b6-4e3d-8960-5b168792a0ba\") " pod="openshift-dns/node-resolver-pmktb" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.253364 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blqh6\" (UniqueName: \"kubernetes.io/projected/134d85a0-99b6-4e3d-8960-5b168792a0ba-kube-api-access-blqh6\") pod \"node-resolver-pmktb\" (UID: \"134d85a0-99b6-4e3d-8960-5b168792a0ba\") " pod="openshift-dns/node-resolver-pmktb" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.289600 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-pmktb" Oct 03 16:24:55 crc kubenswrapper[4744]: W1003 16:24:55.301272 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod134d85a0_99b6_4e3d_8960_5b168792a0ba.slice/crio-5b56e23e8204bea2111c12c8c7eb105443a60b02d1ad0ee561d214fdc92207ac WatchSource:0}: Error finding container 5b56e23e8204bea2111c12c8c7eb105443a60b02d1ad0ee561d214fdc92207ac: Status 404 returned error can't find the container with id 5b56e23e8204bea2111c12c8c7eb105443a60b02d1ad0ee561d214fdc92207ac Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.335225 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.335386 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.335415 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.335543 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.335634 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:24:56.335616358 +0000 UTC m=+22.615492254 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.335690 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.335756 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:24:56.33571346 +0000 UTC m=+22.615589366 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.335803 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:24:56.335792322 +0000 UTC m=+22.615668438 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.436039 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.436088 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.436271 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.436283 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.436337 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.436295 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.436352 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.436368 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.436417 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 16:24:56.436400521 +0000 UTC m=+22.716276407 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.436446 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 16:24:56.436440232 +0000 UTC m=+22.716316128 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.891359 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.891432 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.891543 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:24:55 crc kubenswrapper[4744]: I1003 16:24:55.891454 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.891681 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:24:55 crc kubenswrapper[4744]: E1003 16:24:55.891797 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.019244 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e"} Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.019311 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e"} Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.019323 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cbed9359cb1fdcd4e0a9a586715023420f0b35404ee9125036f2918dfe973c7a"} Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.021282 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4"} Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.021351 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"911125fc0d6d310277cac92a321a9d8428c292f1a2d0f0c941fe40a0ff5a7dff"} Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.027153 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-pmktb" event={"ID":"134d85a0-99b6-4e3d-8960-5b168792a0ba","Type":"ContainerStarted","Data":"0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9"} Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.027217 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-pmktb" event={"ID":"134d85a0-99b6-4e3d-8960-5b168792a0ba","Type":"ContainerStarted","Data":"5b56e23e8204bea2111c12c8c7eb105443a60b02d1ad0ee561d214fdc92207ac"} Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.029458 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d1e2694da0115773f7dac284cb9e8cb9a75bd38930ddfd49da123355c2eeec58"} Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.039697 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.052574 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.067755 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.080827 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.105060 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.120517 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.133040 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.149801 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.165538 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.180060 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.198932 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.214880 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.229169 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.242108 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.254426 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.268587 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.342980 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.343085 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.343146 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:24:56 crc kubenswrapper[4744]: E1003 16:24:56.343272 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:24:56 crc kubenswrapper[4744]: E1003 16:24:56.343554 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:24:58.343489666 +0000 UTC m=+24.623365562 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:24:56 crc kubenswrapper[4744]: E1003 16:24:56.343291 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:24:56 crc kubenswrapper[4744]: E1003 16:24:56.343884 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:24:58.343870535 +0000 UTC m=+24.623746431 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:24:56 crc kubenswrapper[4744]: E1003 16:24:56.343931 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:24:58.343905346 +0000 UTC m=+24.623781232 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.444229 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.444287 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:24:56 crc kubenswrapper[4744]: E1003 16:24:56.444474 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:24:56 crc kubenswrapper[4744]: E1003 16:24:56.444533 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:24:56 crc kubenswrapper[4744]: E1003 16:24:56.444531 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:24:56 crc kubenswrapper[4744]: E1003 16:24:56.444549 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:56 crc kubenswrapper[4744]: E1003 16:24:56.444564 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:24:56 crc kubenswrapper[4744]: E1003 16:24:56.444582 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:56 crc kubenswrapper[4744]: E1003 16:24:56.444627 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 16:24:58.444605766 +0000 UTC m=+24.724481662 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:56 crc kubenswrapper[4744]: E1003 16:24:56.444659 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 16:24:58.444639347 +0000 UTC m=+24.724515443 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.472631 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-vqz6q"] Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.473199 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-qzw8f"] Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.473368 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.473728 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.474881 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-bdrsw"] Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.475646 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.475686 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.476590 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.477121 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.477121 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.477485 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.477770 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.478237 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.478439 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.478863 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.478456 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.488996 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.500798 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.524627 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.544767 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knqt4\" (UniqueName: \"kubernetes.io/projected/6190db0e-c421-41e5-acc2-cf96c34fe3e7-kube-api-access-knqt4\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.545047 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-var-lib-cni-multus\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.545163 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-var-lib-kubelet\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.545266 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6190db0e-c421-41e5-acc2-cf96c34fe3e7-os-release\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.545353 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6190db0e-c421-41e5-acc2-cf96c34fe3e7-cnibin\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.545461 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-run-k8s-cni-cncf-io\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.545576 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cc9bea96-4847-44c8-ab50-d469a6c6155d-mcd-auth-proxy-config\") pod \"machine-config-daemon-qzw8f\" (UID: \"cc9bea96-4847-44c8-ab50-d469a6c6155d\") " pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.545679 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-etc-kubernetes\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.545777 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-multus-cni-dir\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.545864 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-multus-socket-dir-parent\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.545941 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-var-lib-cni-bin\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.546034 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/84192d79-366a-453f-b70b-aefa4537ec4c-multus-daemon-config\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.546129 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqp6h\" (UniqueName: \"kubernetes.io/projected/cc9bea96-4847-44c8-ab50-d469a6c6155d-kube-api-access-rqp6h\") pod \"machine-config-daemon-qzw8f\" (UID: \"cc9bea96-4847-44c8-ab50-d469a6c6155d\") " pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.546211 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6190db0e-c421-41e5-acc2-cf96c34fe3e7-tuning-conf-dir\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.546307 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-system-cni-dir\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.546391 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b757\" (UniqueName: \"kubernetes.io/projected/84192d79-366a-453f-b70b-aefa4537ec4c-kube-api-access-5b757\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.546487 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-hostroot\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.546617 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-multus-conf-dir\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.546718 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cc9bea96-4847-44c8-ab50-d469a6c6155d-proxy-tls\") pod \"machine-config-daemon-qzw8f\" (UID: \"cc9bea96-4847-44c8-ab50-d469a6c6155d\") " pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.546846 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cc9bea96-4847-44c8-ab50-d469a6c6155d-rootfs\") pod \"machine-config-daemon-qzw8f\" (UID: \"cc9bea96-4847-44c8-ab50-d469a6c6155d\") " pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.546947 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6190db0e-c421-41e5-acc2-cf96c34fe3e7-cni-binary-copy\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.547079 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-cnibin\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.547250 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/84192d79-366a-453f-b70b-aefa4537ec4c-cni-binary-copy\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.547318 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6190db0e-c421-41e5-acc2-cf96c34fe3e7-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.547344 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-os-release\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.547385 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6190db0e-c421-41e5-acc2-cf96c34fe3e7-system-cni-dir\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.547405 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-run-netns\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.547426 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-run-multus-certs\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.567792 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.606104 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.628303 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.648435 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-run-multus-certs\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.648752 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knqt4\" (UniqueName: \"kubernetes.io/projected/6190db0e-c421-41e5-acc2-cf96c34fe3e7-kube-api-access-knqt4\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.648838 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-var-lib-cni-multus\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.648924 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-var-lib-kubelet\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.648994 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6190db0e-c421-41e5-acc2-cf96c34fe3e7-os-release\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649045 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-var-lib-kubelet\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.648650 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-run-multus-certs\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649068 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6190db0e-c421-41e5-acc2-cf96c34fe3e7-cnibin\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649177 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-run-k8s-cni-cncf-io\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.648940 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-var-lib-cni-multus\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649244 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cc9bea96-4847-44c8-ab50-d469a6c6155d-mcd-auth-proxy-config\") pod \"machine-config-daemon-qzw8f\" (UID: \"cc9bea96-4847-44c8-ab50-d469a6c6155d\") " pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649310 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-etc-kubernetes\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649370 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-multus-cni-dir\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649392 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-multus-socket-dir-parent\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649403 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-etc-kubernetes\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649477 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6190db0e-c421-41e5-acc2-cf96c34fe3e7-os-release\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649485 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-multus-socket-dir-parent\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649532 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-run-k8s-cni-cncf-io\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649556 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-var-lib-cni-bin\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649325 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6190db0e-c421-41e5-acc2-cf96c34fe3e7-cnibin\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649658 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-multus-cni-dir\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649700 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-var-lib-cni-bin\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649725 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/84192d79-366a-453f-b70b-aefa4537ec4c-multus-daemon-config\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649749 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqp6h\" (UniqueName: \"kubernetes.io/projected/cc9bea96-4847-44c8-ab50-d469a6c6155d-kube-api-access-rqp6h\") pod \"machine-config-daemon-qzw8f\" (UID: \"cc9bea96-4847-44c8-ab50-d469a6c6155d\") " pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649769 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6190db0e-c421-41e5-acc2-cf96c34fe3e7-tuning-conf-dir\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649787 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-system-cni-dir\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649806 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b757\" (UniqueName: \"kubernetes.io/projected/84192d79-366a-453f-b70b-aefa4537ec4c-kube-api-access-5b757\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649825 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-hostroot\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649858 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-multus-conf-dir\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649876 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cc9bea96-4847-44c8-ab50-d469a6c6155d-proxy-tls\") pod \"machine-config-daemon-qzw8f\" (UID: \"cc9bea96-4847-44c8-ab50-d469a6c6155d\") " pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cc9bea96-4847-44c8-ab50-d469a6c6155d-rootfs\") pod \"machine-config-daemon-qzw8f\" (UID: \"cc9bea96-4847-44c8-ab50-d469a6c6155d\") " pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649937 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6190db0e-c421-41e5-acc2-cf96c34fe3e7-cni-binary-copy\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649955 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-cnibin\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649973 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/84192d79-366a-453f-b70b-aefa4537ec4c-cni-binary-copy\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.649995 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6190db0e-c421-41e5-acc2-cf96c34fe3e7-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.650014 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-os-release\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.650046 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6190db0e-c421-41e5-acc2-cf96c34fe3e7-system-cni-dir\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.650063 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-run-netns\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.650093 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cc9bea96-4847-44c8-ab50-d469a6c6155d-mcd-auth-proxy-config\") pod \"machine-config-daemon-qzw8f\" (UID: \"cc9bea96-4847-44c8-ab50-d469a6c6155d\") " pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.650118 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-host-run-netns\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.650151 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cc9bea96-4847-44c8-ab50-d469a6c6155d-rootfs\") pod \"machine-config-daemon-qzw8f\" (UID: \"cc9bea96-4847-44c8-ab50-d469a6c6155d\") " pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.650353 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-hostroot\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.650406 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-multus-conf-dir\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.650360 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-system-cni-dir\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.650479 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-cnibin\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.650517 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6190db0e-c421-41e5-acc2-cf96c34fe3e7-system-cni-dir\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.650706 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/84192d79-366a-453f-b70b-aefa4537ec4c-os-release\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.650894 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.651050 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/6190db0e-c421-41e5-acc2-cf96c34fe3e7-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.651062 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6190db0e-c421-41e5-acc2-cf96c34fe3e7-tuning-conf-dir\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.651606 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/84192d79-366a-453f-b70b-aefa4537ec4c-multus-daemon-config\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.651628 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6190db0e-c421-41e5-acc2-cf96c34fe3e7-cni-binary-copy\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.651767 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/84192d79-366a-453f-b70b-aefa4537ec4c-cni-binary-copy\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.660983 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cc9bea96-4847-44c8-ab50-d469a6c6155d-proxy-tls\") pod \"machine-config-daemon-qzw8f\" (UID: \"cc9bea96-4847-44c8-ab50-d469a6c6155d\") " pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.669754 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqp6h\" (UniqueName: \"kubernetes.io/projected/cc9bea96-4847-44c8-ab50-d469a6c6155d-kube-api-access-rqp6h\") pod \"machine-config-daemon-qzw8f\" (UID: \"cc9bea96-4847-44c8-ab50-d469a6c6155d\") " pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.670957 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.671875 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b757\" (UniqueName: \"kubernetes.io/projected/84192d79-366a-453f-b70b-aefa4537ec4c-kube-api-access-5b757\") pod \"multus-vqz6q\" (UID: \"84192d79-366a-453f-b70b-aefa4537ec4c\") " pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.672249 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knqt4\" (UniqueName: \"kubernetes.io/projected/6190db0e-c421-41e5-acc2-cf96c34fe3e7-kube-api-access-knqt4\") pod \"multus-additional-cni-plugins-bdrsw\" (UID: \"6190db0e-c421-41e5-acc2-cf96c34fe3e7\") " pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.687220 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.699360 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.713320 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.732182 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.746471 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.763548 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.776265 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.789628 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.794029 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vqz6q" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.802619 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.804231 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: W1003 16:24:56.807112 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84192d79_366a_453f_b70b_aefa4537ec4c.slice/crio-8cd28260ab7c1bd2feebf6f441353a4aaa506e48be5c0dd3fd206c38af52412b WatchSource:0}: Error finding container 8cd28260ab7c1bd2feebf6f441353a4aaa506e48be5c0dd3fd206c38af52412b: Status 404 returned error can't find the container with id 8cd28260ab7c1bd2feebf6f441353a4aaa506e48be5c0dd3fd206c38af52412b Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.812452 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.815965 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: W1003 16:24:56.817632 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc9bea96_4847_44c8_ab50_d469a6c6155d.slice/crio-431f7c4244f8c66cc07fc8c647ee56d53bd55612846ac13420067f52f056bfdd WatchSource:0}: Error finding container 431f7c4244f8c66cc07fc8c647ee56d53bd55612846ac13420067f52f056bfdd: Status 404 returned error can't find the container with id 431f7c4244f8c66cc07fc8c647ee56d53bd55612846ac13420067f52f056bfdd Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.842820 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.862204 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.877394 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.898862 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jw5rc"] Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.899800 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.904080 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.904371 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.904426 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.904539 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.904546 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.904680 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.904812 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.906719 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.921427 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.949514 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.965700 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.981773 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:56 crc kubenswrapper[4744]: I1003 16:24:56.998447 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:56Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.010966 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.031597 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.038541 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" event={"ID":"6190db0e-c421-41e5-acc2-cf96c34fe3e7","Type":"ContainerStarted","Data":"fcce079769b52ed61e20a927439661ce884f3eeac7209fad30f324ef88ecae67"} Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.040543 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"431f7c4244f8c66cc07fc8c647ee56d53bd55612846ac13420067f52f056bfdd"} Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.044292 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vqz6q" event={"ID":"84192d79-366a-453f-b70b-aefa4537ec4c","Type":"ContainerStarted","Data":"8cd28260ab7c1bd2feebf6f441353a4aaa506e48be5c0dd3fd206c38af52412b"} Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.049841 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.053748 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-systemd-units\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.053790 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-openvswitch\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.053816 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-cni-netd\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.053834 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovnkube-config\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.053940 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-slash\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.053982 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxw7s\" (UniqueName: \"kubernetes.io/projected/3436425f-b9cf-4302-b5ad-ccb92d597aed-kube-api-access-nxw7s\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.054013 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-log-socket\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.054035 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-run-ovn-kubernetes\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.054055 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-node-log\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.054157 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-ovn\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.054195 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovn-node-metrics-cert\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.054222 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovnkube-script-lib\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.054248 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-etc-openvswitch\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.054276 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-systemd\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.054319 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-kubelet\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.054336 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-run-netns\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.054352 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-var-lib-openvswitch\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.054379 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-cni-bin\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.054395 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.054417 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-env-overrides\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.088177 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.105542 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.128155 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.142346 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156032 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156084 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-env-overrides\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156128 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-cni-bin\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156150 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-systemd-units\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156200 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-openvswitch\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156225 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-cni-netd\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156236 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-cni-bin\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156246 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovnkube-config\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156348 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-slash\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156374 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxw7s\" (UniqueName: \"kubernetes.io/projected/3436425f-b9cf-4302-b5ad-ccb92d597aed-kube-api-access-nxw7s\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156409 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-log-socket\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156432 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-run-ovn-kubernetes\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156452 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-node-log\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156473 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-ovn\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156515 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovn-node-metrics-cert\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156535 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovnkube-script-lib\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156571 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-etc-openvswitch\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156594 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-systemd\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156621 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-kubelet\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156640 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-run-netns\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156658 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-var-lib-openvswitch\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156728 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-var-lib-openvswitch\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156752 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-slash\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156935 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-env-overrides\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.157017 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovnkube-config\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.157081 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-systemd-units\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.157116 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-openvswitch\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.156191 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.157156 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-cni-netd\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.157183 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-systemd\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.157211 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-etc-openvswitch\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.157236 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-run-ovn-kubernetes\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.157265 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-log-socket\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.157290 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-node-log\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.157314 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-kubelet\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.157339 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-run-netns\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.157363 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-ovn\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.157748 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovnkube-script-lib\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.164110 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovn-node-metrics-cert\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.176057 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxw7s\" (UniqueName: \"kubernetes.io/projected/3436425f-b9cf-4302-b5ad-ccb92d597aed-kube-api-access-nxw7s\") pod \"ovnkube-node-jw5rc\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.228983 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:24:57 crc kubenswrapper[4744]: W1003 16:24:57.241334 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3436425f_b9cf_4302_b5ad_ccb92d597aed.slice/crio-729b5896f0a8615cceb963cc8a7da51f07b98eaad1f1e4c70ded6c519d99f2f6 WatchSource:0}: Error finding container 729b5896f0a8615cceb963cc8a7da51f07b98eaad1f1e4c70ded6c519d99f2f6: Status 404 returned error can't find the container with id 729b5896f0a8615cceb963cc8a7da51f07b98eaad1f1e4c70ded6c519d99f2f6 Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.351716 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.365926 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.371484 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.373528 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.387452 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.403241 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.418640 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.434285 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.448849 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.463396 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.475321 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.494352 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.529564 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.546999 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.564851 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.592388 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.607447 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.624605 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.639169 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.654378 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.666890 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.681775 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.697280 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.714615 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.740599 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.756127 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.772280 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.789708 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:57Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.891789 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.891923 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:24:57 crc kubenswrapper[4744]: I1003 16:24:57.892092 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:24:57 crc kubenswrapper[4744]: E1003 16:24:57.892370 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:24:57 crc kubenswrapper[4744]: E1003 16:24:57.891949 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:24:57 crc kubenswrapper[4744]: E1003 16:24:57.892113 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.049570 4744 generic.go:334] "Generic (PLEG): container finished" podID="6190db0e-c421-41e5-acc2-cf96c34fe3e7" containerID="4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5" exitCode=0 Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.049688 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" event={"ID":"6190db0e-c421-41e5-acc2-cf96c34fe3e7","Type":"ContainerDied","Data":"4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5"} Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.051140 4744 generic.go:334] "Generic (PLEG): container finished" podID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerID="a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b" exitCode=0 Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.051209 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerDied","Data":"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b"} Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.051227 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerStarted","Data":"729b5896f0a8615cceb963cc8a7da51f07b98eaad1f1e4c70ded6c519d99f2f6"} Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.053426 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11"} Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.055640 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d"} Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.055688 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e"} Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.058144 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vqz6q" event={"ID":"84192d79-366a-453f-b70b-aefa4537ec4c","Type":"ContainerStarted","Data":"c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17"} Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.073804 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.096615 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.120627 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.141056 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.141629 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.146831 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.151700 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.163283 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.178999 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.192109 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.212923 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.227643 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.245574 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.259706 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.276056 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.291478 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.306627 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.321046 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.335670 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.350783 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.363383 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.369212 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:24:58 crc kubenswrapper[4744]: E1003 16:24:58.369337 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:25:02.369307522 +0000 UTC m=+28.649183418 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.369389 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.369465 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:24:58 crc kubenswrapper[4744]: E1003 16:24:58.369572 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:24:58 crc kubenswrapper[4744]: E1003 16:24:58.369617 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:24:58 crc kubenswrapper[4744]: E1003 16:24:58.369645 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:02.36962569 +0000 UTC m=+28.649501586 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:24:58 crc kubenswrapper[4744]: E1003 16:24:58.369666 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:02.36965878 +0000 UTC m=+28.649534676 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.382287 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.398190 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.412072 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.427545 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.427738 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-6rjqm"] Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.428151 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-6rjqm" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.429780 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.430271 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.430558 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.432479 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.455990 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.470706 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.470767 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:24:58 crc kubenswrapper[4744]: E1003 16:24:58.470911 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:24:58 crc kubenswrapper[4744]: E1003 16:24:58.470927 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:24:58 crc kubenswrapper[4744]: E1003 16:24:58.470930 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:24:58 crc kubenswrapper[4744]: E1003 16:24:58.470979 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:24:58 crc kubenswrapper[4744]: E1003 16:24:58.470992 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:58 crc kubenswrapper[4744]: E1003 16:24:58.471065 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:02.471041947 +0000 UTC m=+28.750917843 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:58 crc kubenswrapper[4744]: E1003 16:24:58.470940 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:58 crc kubenswrapper[4744]: E1003 16:24:58.471113 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:02.471107609 +0000 UTC m=+28.750983505 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.499698 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.529796 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.570369 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.571730 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e799ae45-a110-4d69-8c66-3d89111eab93-host\") pod \"node-ca-6rjqm\" (UID: \"e799ae45-a110-4d69-8c66-3d89111eab93\") " pod="openshift-image-registry/node-ca-6rjqm" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.571834 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e799ae45-a110-4d69-8c66-3d89111eab93-serviceca\") pod \"node-ca-6rjqm\" (UID: \"e799ae45-a110-4d69-8c66-3d89111eab93\") " pod="openshift-image-registry/node-ca-6rjqm" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.571891 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd2s8\" (UniqueName: \"kubernetes.io/projected/e799ae45-a110-4d69-8c66-3d89111eab93-kube-api-access-pd2s8\") pod \"node-ca-6rjqm\" (UID: \"e799ae45-a110-4d69-8c66-3d89111eab93\") " pod="openshift-image-registry/node-ca-6rjqm" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.612766 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.651245 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.673456 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd2s8\" (UniqueName: \"kubernetes.io/projected/e799ae45-a110-4d69-8c66-3d89111eab93-kube-api-access-pd2s8\") pod \"node-ca-6rjqm\" (UID: \"e799ae45-a110-4d69-8c66-3d89111eab93\") " pod="openshift-image-registry/node-ca-6rjqm" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.673570 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e799ae45-a110-4d69-8c66-3d89111eab93-host\") pod \"node-ca-6rjqm\" (UID: \"e799ae45-a110-4d69-8c66-3d89111eab93\") " pod="openshift-image-registry/node-ca-6rjqm" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.673607 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e799ae45-a110-4d69-8c66-3d89111eab93-serviceca\") pod \"node-ca-6rjqm\" (UID: \"e799ae45-a110-4d69-8c66-3d89111eab93\") " pod="openshift-image-registry/node-ca-6rjqm" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.673717 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e799ae45-a110-4d69-8c66-3d89111eab93-host\") pod \"node-ca-6rjqm\" (UID: \"e799ae45-a110-4d69-8c66-3d89111eab93\") " pod="openshift-image-registry/node-ca-6rjqm" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.676612 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e799ae45-a110-4d69-8c66-3d89111eab93-serviceca\") pod \"node-ca-6rjqm\" (UID: \"e799ae45-a110-4d69-8c66-3d89111eab93\") " pod="openshift-image-registry/node-ca-6rjqm" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.689233 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.718808 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd2s8\" (UniqueName: \"kubernetes.io/projected/e799ae45-a110-4d69-8c66-3d89111eab93-kube-api-access-pd2s8\") pod \"node-ca-6rjqm\" (UID: \"e799ae45-a110-4d69-8c66-3d89111eab93\") " pod="openshift-image-registry/node-ca-6rjqm" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.762847 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.803366 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.836149 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.871966 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.908728 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.952036 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.960092 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-6rjqm" Oct 03 16:24:58 crc kubenswrapper[4744]: W1003 16:24:58.989455 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode799ae45_a110_4d69_8c66_3d89111eab93.slice/crio-61a982615f531f6bda0bf07ee4f8bf46aa4da53e46a7869d60b04dc09b68ac07 WatchSource:0}: Error finding container 61a982615f531f6bda0bf07ee4f8bf46aa4da53e46a7869d60b04dc09b68ac07: Status 404 returned error can't find the container with id 61a982615f531f6bda0bf07ee4f8bf46aa4da53e46a7869d60b04dc09b68ac07 Oct 03 16:24:58 crc kubenswrapper[4744]: I1003 16:24:58.990448 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.030778 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.067288 4744 generic.go:334] "Generic (PLEG): container finished" podID="6190db0e-c421-41e5-acc2-cf96c34fe3e7" containerID="f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505" exitCode=0 Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.068323 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" event={"ID":"6190db0e-c421-41e5-acc2-cf96c34fe3e7","Type":"ContainerDied","Data":"f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505"} Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.072832 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.076580 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerStarted","Data":"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c"} Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.076621 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerStarted","Data":"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603"} Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.076632 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerStarted","Data":"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558"} Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.076641 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerStarted","Data":"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0"} Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.078465 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-6rjqm" event={"ID":"e799ae45-a110-4d69-8c66-3d89111eab93","Type":"ContainerStarted","Data":"61a982615f531f6bda0bf07ee4f8bf46aa4da53e46a7869d60b04dc09b68ac07"} Oct 03 16:24:59 crc kubenswrapper[4744]: E1003 16:24:59.106728 4744 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.132341 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.164614 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.179581 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.216448 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.251000 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.292033 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.330382 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.370307 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.411946 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.448926 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.499221 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.530529 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.570648 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.614766 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.657037 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.691764 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.732372 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.778135 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.812303 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.850119 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:24:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.890921 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.891410 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:24:59 crc kubenswrapper[4744]: E1003 16:24:59.891538 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:24:59 crc kubenswrapper[4744]: I1003 16:24:59.891604 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:24:59 crc kubenswrapper[4744]: E1003 16:24:59.891748 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:24:59 crc kubenswrapper[4744]: E1003 16:24:59.891890 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.086427 4744 generic.go:334] "Generic (PLEG): container finished" podID="6190db0e-c421-41e5-acc2-cf96c34fe3e7" containerID="6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037" exitCode=0 Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.086486 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" event={"ID":"6190db0e-c421-41e5-acc2-cf96c34fe3e7","Type":"ContainerDied","Data":"6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037"} Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.094111 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerStarted","Data":"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64"} Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.094162 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerStarted","Data":"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d"} Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.097447 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-6rjqm" event={"ID":"e799ae45-a110-4d69-8c66-3d89111eab93","Type":"ContainerStarted","Data":"ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f"} Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.112533 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.130195 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.147275 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.163824 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.178302 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.202158 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.216194 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.229827 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.244272 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.255999 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.291229 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.338979 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.369454 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.410156 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.451793 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.498367 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.533472 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.568995 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.611020 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.663341 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.693795 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.731645 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.771994 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.814556 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.854488 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.891601 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.931841 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:00 crc kubenswrapper[4744]: I1003 16:25:00.970875 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.011775 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.049561 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.102360 4744 generic.go:334] "Generic (PLEG): container finished" podID="6190db0e-c421-41e5-acc2-cf96c34fe3e7" containerID="783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a" exitCode=0 Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.102422 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" event={"ID":"6190db0e-c421-41e5-acc2-cf96c34fe3e7","Type":"ContainerDied","Data":"783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a"} Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.118791 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.121989 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.122055 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.122080 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.122252 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.141804 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.149767 4744 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.150098 4744 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.151482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.151540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.151550 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.151570 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.151578 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:01Z","lastTransitionTime":"2025-10-03T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:01 crc kubenswrapper[4744]: E1003 16:25:01.176104 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.182708 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.182752 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.182761 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.182776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.182786 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:01Z","lastTransitionTime":"2025-10-03T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.195827 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: E1003 16:25:01.202967 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.206913 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.206966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.206978 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.206996 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.207007 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:01Z","lastTransitionTime":"2025-10-03T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.212023 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: E1003 16:25:01.221051 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.225289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.225353 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.225384 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.225405 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.225417 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:01Z","lastTransitionTime":"2025-10-03T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:01 crc kubenswrapper[4744]: E1003 16:25:01.237470 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.241806 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.241834 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.241845 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.241860 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.241869 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:01Z","lastTransitionTime":"2025-10-03T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.250298 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: E1003 16:25:01.256754 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: E1003 16:25:01.256865 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.260982 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.261013 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.261027 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.261047 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.261059 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:01Z","lastTransitionTime":"2025-10-03T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.292314 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.329611 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.363847 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.363892 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.363906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.363924 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.363936 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:01Z","lastTransitionTime":"2025-10-03T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.370523 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.411748 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.450292 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.466455 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.466811 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.466821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.466837 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.466848 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:01Z","lastTransitionTime":"2025-10-03T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.488565 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.529611 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.568740 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.568790 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.568799 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.568815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.568824 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:01Z","lastTransitionTime":"2025-10-03T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.576257 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.610051 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.653486 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.670861 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.670897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.670907 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.670923 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.670933 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:01Z","lastTransitionTime":"2025-10-03T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.693661 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:01Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.772816 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.772864 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.772874 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.772892 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.772903 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:01Z","lastTransitionTime":"2025-10-03T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.875706 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.875755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.875766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.875784 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.875797 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:01Z","lastTransitionTime":"2025-10-03T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.891098 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.891109 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.891230 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:01 crc kubenswrapper[4744]: E1003 16:25:01.891279 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:01 crc kubenswrapper[4744]: E1003 16:25:01.891560 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:01 crc kubenswrapper[4744]: E1003 16:25:01.891671 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.979117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.979189 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.979203 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.979227 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:01 crc kubenswrapper[4744]: I1003 16:25:01.979249 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:01Z","lastTransitionTime":"2025-10-03T16:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.081776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.081808 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.081816 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.081832 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.081840 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:02Z","lastTransitionTime":"2025-10-03T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.109257 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerStarted","Data":"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe"} Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.112743 4744 generic.go:334] "Generic (PLEG): container finished" podID="6190db0e-c421-41e5-acc2-cf96c34fe3e7" containerID="1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c" exitCode=0 Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.112837 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" event={"ID":"6190db0e-c421-41e5-acc2-cf96c34fe3e7","Type":"ContainerDied","Data":"1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c"} Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.137860 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:02Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.151425 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:02Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.163301 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:02Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.181924 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:02Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.185226 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.185279 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.185302 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.185325 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.185336 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:02Z","lastTransitionTime":"2025-10-03T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.203227 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:02Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.217804 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:02Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.232766 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:02Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.244455 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:02Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.259400 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:02Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.271586 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:02Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.287828 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.287879 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.287893 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.287914 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.287927 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:02Z","lastTransitionTime":"2025-10-03T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.288096 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:02Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.301572 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:02Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.314285 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:02Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.326654 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:02Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.336003 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:02Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.390306 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.390351 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.390364 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.390381 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.390393 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:02Z","lastTransitionTime":"2025-10-03T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.412741 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.412892 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:02 crc kubenswrapper[4744]: E1003 16:25:02.412924 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:25:10.412900742 +0000 UTC m=+36.692776638 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:25:02 crc kubenswrapper[4744]: E1003 16:25:02.412971 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.412988 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:02 crc kubenswrapper[4744]: E1003 16:25:02.413006 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:10.412997964 +0000 UTC m=+36.692873850 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:25:02 crc kubenswrapper[4744]: E1003 16:25:02.413085 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:25:02 crc kubenswrapper[4744]: E1003 16:25:02.413126 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:10.413116197 +0000 UTC m=+36.692992093 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.493527 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.493568 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.493580 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.493606 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.493618 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:02Z","lastTransitionTime":"2025-10-03T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.513646 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.513707 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:02 crc kubenswrapper[4744]: E1003 16:25:02.513826 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:25:02 crc kubenswrapper[4744]: E1003 16:25:02.513828 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:25:02 crc kubenswrapper[4744]: E1003 16:25:02.513853 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:25:02 crc kubenswrapper[4744]: E1003 16:25:02.513868 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:25:02 crc kubenswrapper[4744]: E1003 16:25:02.513872 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:02 crc kubenswrapper[4744]: E1003 16:25:02.513886 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:02 crc kubenswrapper[4744]: E1003 16:25:02.513941 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:10.51392015 +0000 UTC m=+36.793796106 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:02 crc kubenswrapper[4744]: E1003 16:25:02.513960 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:10.513951861 +0000 UTC m=+36.793827857 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.596933 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.597006 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.597031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.597059 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.597079 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:02Z","lastTransitionTime":"2025-10-03T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.700214 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.700279 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.700299 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.700326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.700346 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:02Z","lastTransitionTime":"2025-10-03T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.803140 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.803194 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.803206 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.803223 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.803237 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:02Z","lastTransitionTime":"2025-10-03T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.905730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.905769 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.905780 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.905794 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:02 crc kubenswrapper[4744]: I1003 16:25:02.905804 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:02Z","lastTransitionTime":"2025-10-03T16:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.007807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.007844 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.007853 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.007869 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.007878 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:03Z","lastTransitionTime":"2025-10-03T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.110798 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.110845 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.110854 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.110870 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.110879 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:03Z","lastTransitionTime":"2025-10-03T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.119015 4744 generic.go:334] "Generic (PLEG): container finished" podID="6190db0e-c421-41e5-acc2-cf96c34fe3e7" containerID="4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583" exitCode=0 Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.119071 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" event={"ID":"6190db0e-c421-41e5-acc2-cf96c34fe3e7","Type":"ContainerDied","Data":"4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583"} Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.134857 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.153117 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.164179 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.182284 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.194698 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.208829 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.213661 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.213708 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.213719 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.213737 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.213749 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:03Z","lastTransitionTime":"2025-10-03T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.219485 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.229270 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.243068 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.256973 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.270650 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.281989 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.296770 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.316732 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.316772 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.316782 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.316797 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.316808 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:03Z","lastTransitionTime":"2025-10-03T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.319018 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.339555 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.420111 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.420161 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.420174 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.420192 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.420203 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:03Z","lastTransitionTime":"2025-10-03T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.522564 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.523129 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.523144 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.523165 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.523180 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:03Z","lastTransitionTime":"2025-10-03T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.625775 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.625841 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.625868 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.625912 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.625938 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:03Z","lastTransitionTime":"2025-10-03T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.728947 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.729009 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.729022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.729041 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.729053 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:03Z","lastTransitionTime":"2025-10-03T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.831616 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.831749 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.831776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.831806 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.831822 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:03Z","lastTransitionTime":"2025-10-03T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.891852 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.891847 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.891867 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:03 crc kubenswrapper[4744]: E1003 16:25:03.891982 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:03 crc kubenswrapper[4744]: E1003 16:25:03.892185 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:03 crc kubenswrapper[4744]: E1003 16:25:03.892256 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.934003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.934039 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.934048 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.934064 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:03 crc kubenswrapper[4744]: I1003 16:25:03.934072 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:03Z","lastTransitionTime":"2025-10-03T16:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.036513 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.036551 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.036559 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.036573 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.036583 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:04Z","lastTransitionTime":"2025-10-03T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.127577 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerStarted","Data":"f5b67e587aaf68f1c700802f2ee729e3b8f4bb791de08992d83a5afa0cab2f7c"} Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.127782 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.132485 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" event={"ID":"6190db0e-c421-41e5-acc2-cf96c34fe3e7","Type":"ContainerStarted","Data":"6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44"} Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.139302 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.139342 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.139352 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.139367 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.139382 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:04Z","lastTransitionTime":"2025-10-03T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.146926 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.160043 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.165854 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.177977 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.195168 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.211643 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.224530 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.242902 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.243403 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.243642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.243833 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.243993 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:04Z","lastTransitionTime":"2025-10-03T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.249941 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.284330 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.298029 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.313463 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.328551 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.344486 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.346456 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.346525 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.346541 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.346564 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.346579 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:04Z","lastTransitionTime":"2025-10-03T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.364835 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b67e587aaf68f1c700802f2ee729e3b8f4bb791de08992d83a5afa0cab2f7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.378158 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.392757 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.404446 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.415857 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.432452 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.449876 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.449926 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.449937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.449953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.449963 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:04Z","lastTransitionTime":"2025-10-03T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.453398 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.466809 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.479066 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.491295 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.504178 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.524430 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b67e587aaf68f1c700802f2ee729e3b8f4bb791de08992d83a5afa0cab2f7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.537060 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.552913 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.552997 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.553012 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.553037 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.553058 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:04Z","lastTransitionTime":"2025-10-03T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.553134 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.565221 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.576460 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.590869 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.604137 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.655214 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.655260 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.655268 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.655282 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.655294 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:04Z","lastTransitionTime":"2025-10-03T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.757436 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.757489 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.757518 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.757533 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.757546 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:04Z","lastTransitionTime":"2025-10-03T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.860031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.860068 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.860077 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.860092 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.860105 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:04Z","lastTransitionTime":"2025-10-03T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.907845 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.945056 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.962612 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.962650 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.962662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.962680 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.962693 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:04Z","lastTransitionTime":"2025-10-03T16:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.967553 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:04 crc kubenswrapper[4744]: I1003 16:25:04.997416 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.016759 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.030730 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.051303 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b67e587aaf68f1c700802f2ee729e3b8f4bb791de08992d83a5afa0cab2f7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.065161 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.065223 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.065234 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.065253 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.065263 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:05Z","lastTransitionTime":"2025-10-03T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.070863 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.088916 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.103797 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.117256 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.134313 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.136141 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.137810 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.155833 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.162298 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.168044 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.168090 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.168109 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.168140 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.168164 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:05Z","lastTransitionTime":"2025-10-03T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.171146 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.183052 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.202179 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.214907 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.227097 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.240634 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.252941 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.267919 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.270416 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.270476 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.270517 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.270540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.270554 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:05Z","lastTransitionTime":"2025-10-03T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.284392 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.298730 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.319171 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.339455 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b67e587aaf68f1c700802f2ee729e3b8f4bb791de08992d83a5afa0cab2f7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.356322 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.370906 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.372602 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.372642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.372663 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.372681 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.372691 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:05Z","lastTransitionTime":"2025-10-03T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.380992 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.392030 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.405611 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.475350 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.475400 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.475412 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.475430 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.475445 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:05Z","lastTransitionTime":"2025-10-03T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.581964 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.582026 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.582042 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.582063 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.582076 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:05Z","lastTransitionTime":"2025-10-03T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.684432 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.684463 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.684471 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.684485 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.684512 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:05Z","lastTransitionTime":"2025-10-03T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.787571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.787610 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.787622 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.787639 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.787649 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:05Z","lastTransitionTime":"2025-10-03T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.889403 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.889434 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.889443 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.889456 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.889464 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:05Z","lastTransitionTime":"2025-10-03T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.891086 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:05 crc kubenswrapper[4744]: E1003 16:25:05.891174 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.891460 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:05 crc kubenswrapper[4744]: E1003 16:25:05.891529 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.891567 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:05 crc kubenswrapper[4744]: E1003 16:25:05.891606 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.991939 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.991997 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.992009 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.992031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:05 crc kubenswrapper[4744]: I1003 16:25:05.992044 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:05Z","lastTransitionTime":"2025-10-03T16:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.094861 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.094900 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.094908 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.094925 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.094935 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:06Z","lastTransitionTime":"2025-10-03T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.139352 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.197240 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.197289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.197299 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.197314 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.197324 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:06Z","lastTransitionTime":"2025-10-03T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.299947 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.299999 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.300007 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.300022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.300048 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:06Z","lastTransitionTime":"2025-10-03T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.403153 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.403198 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.403207 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.403232 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.403242 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:06Z","lastTransitionTime":"2025-10-03T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.505759 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.505840 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.505851 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.505871 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.505884 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:06Z","lastTransitionTime":"2025-10-03T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.608313 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.608364 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.608374 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.608390 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.608402 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:06Z","lastTransitionTime":"2025-10-03T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.710279 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.710359 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.710388 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.710418 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.710439 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:06Z","lastTransitionTime":"2025-10-03T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.813190 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.813228 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.813236 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.813253 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.813262 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:06Z","lastTransitionTime":"2025-10-03T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.915213 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.915270 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.915282 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.915304 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:06 crc kubenswrapper[4744]: I1003 16:25:06.915318 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:06Z","lastTransitionTime":"2025-10-03T16:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.018781 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.018840 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.018856 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.018885 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.018896 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:07Z","lastTransitionTime":"2025-10-03T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.120892 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.120960 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.120971 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.120987 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.120996 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:07Z","lastTransitionTime":"2025-10-03T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.143551 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovnkube-controller/0.log" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.145996 4744 generic.go:334] "Generic (PLEG): container finished" podID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerID="f5b67e587aaf68f1c700802f2ee729e3b8f4bb791de08992d83a5afa0cab2f7c" exitCode=1 Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.146031 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerDied","Data":"f5b67e587aaf68f1c700802f2ee729e3b8f4bb791de08992d83a5afa0cab2f7c"} Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.146749 4744 scope.go:117] "RemoveContainer" containerID="f5b67e587aaf68f1c700802f2ee729e3b8f4bb791de08992d83a5afa0cab2f7c" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.160097 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:07Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.174896 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:07Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.193952 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:07Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.211513 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:07Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.223899 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.223953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.223965 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.223986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.224242 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:07Z","lastTransitionTime":"2025-10-03T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.238865 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:07Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.259952 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:07Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.274652 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:07Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.288926 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:07Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.304544 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:07Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.317564 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:07Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.326689 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.326721 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.326730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.326747 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.326758 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:07Z","lastTransitionTime":"2025-10-03T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.338809 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b67e587aaf68f1c700802f2ee729e3b8f4bb791de08992d83a5afa0cab2f7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5b67e587aaf68f1c700802f2ee729e3b8f4bb791de08992d83a5afa0cab2f7c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:06Z\\\",\\\"message\\\":\\\"ers/factory.go:160\\\\nI1003 16:25:06.385860 6031 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:06.386051 6031 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:06.386216 6031 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:06.386312 6031 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 16:25:06.386395 6031 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1003 16:25:06.386431 6031 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1003 16:25:06.386459 6031 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1003 16:25:06.386512 6031 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 16:25:06.386564 6031 factory.go:656] Stopping watch factory\\\\nI1003 16:25:06.386586 6031 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1003 16:25:06.386649 6031 ovnkube.go:599] Stopped ovnkube\\\\nI1003 16:25:06.386466 6031 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 16:25:06.386470 6031 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 16:25:06.386522 6031 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:07Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.356585 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:07Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.375869 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:07Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.390783 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:07Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.405282 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:07Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.429698 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.429741 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.429750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.429765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.429775 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:07Z","lastTransitionTime":"2025-10-03T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.532130 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.532172 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.532182 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.532197 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.532208 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:07Z","lastTransitionTime":"2025-10-03T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.634321 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.634380 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.634389 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.634402 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.634413 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:07Z","lastTransitionTime":"2025-10-03T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.736799 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.736840 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.736854 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.736870 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.736887 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:07Z","lastTransitionTime":"2025-10-03T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.839189 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.839228 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.839237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.839253 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.839266 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:07Z","lastTransitionTime":"2025-10-03T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.890927 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.890958 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.891059 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:07 crc kubenswrapper[4744]: E1003 16:25:07.891074 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:07 crc kubenswrapper[4744]: E1003 16:25:07.891119 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:07 crc kubenswrapper[4744]: E1003 16:25:07.891176 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.941751 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.941792 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.941801 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.941818 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:07 crc kubenswrapper[4744]: I1003 16:25:07.941829 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:07Z","lastTransitionTime":"2025-10-03T16:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.044188 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.044235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.044244 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.044260 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.044269 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:08Z","lastTransitionTime":"2025-10-03T16:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.147142 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.147174 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.147214 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.147234 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.147244 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:08Z","lastTransitionTime":"2025-10-03T16:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.149926 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovnkube-controller/1.log" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.150316 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovnkube-controller/0.log" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.152596 4744 generic.go:334] "Generic (PLEG): container finished" podID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerID="9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb" exitCode=1 Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.152634 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerDied","Data":"9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb"} Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.152682 4744 scope.go:117] "RemoveContainer" containerID="f5b67e587aaf68f1c700802f2ee729e3b8f4bb791de08992d83a5afa0cab2f7c" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.153319 4744 scope.go:117] "RemoveContainer" containerID="9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb" Oct 03 16:25:08 crc kubenswrapper[4744]: E1003 16:25:08.153485 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.180417 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.198613 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.212430 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.226225 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.241017 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.249324 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.249359 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.249371 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.249389 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.249401 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:08Z","lastTransitionTime":"2025-10-03T16:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.252512 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.273246 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5b67e587aaf68f1c700802f2ee729e3b8f4bb791de08992d83a5afa0cab2f7c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:06Z\\\",\\\"message\\\":\\\"ers/factory.go:160\\\\nI1003 16:25:06.385860 6031 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:06.386051 6031 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:06.386216 6031 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:06.386312 6031 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 16:25:06.386395 6031 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1003 16:25:06.386431 6031 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1003 16:25:06.386459 6031 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1003 16:25:06.386512 6031 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 16:25:06.386564 6031 factory.go:656] Stopping watch factory\\\\nI1003 16:25:06.386586 6031 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1003 16:25:06.386649 6031 ovnkube.go:599] Stopped ovnkube\\\\nI1003 16:25:06.386466 6031 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 16:25:06.386470 6031 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 16:25:06.386522 6031 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:07Z\\\",\\\"message\\\":\\\"0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:07.869637 6152 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 16:25:07.870150 6152 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1003 16:25:07.870170 6152 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1003 16:25:07.870193 6152 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 16:25:07.870214 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1003 16:25:07.870222 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1003 16:25:07.870247 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 16:25:07.870265 6152 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1003 16:25:07.870276 6152 factory.go:656] Stopping watch factory\\\\nI1003 16:25:07.870279 6152 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 16:25:07.870278 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 16:25:07.870289 6152 ovnkube.go:599] Stopped ovnkube\\\\nI1003 16:25:07.870288 6152 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 16:25:07.870294 6152 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.285989 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.298809 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.311247 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.321110 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.337603 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.350291 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.351883 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.351917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.351934 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.351956 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.351970 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:08Z","lastTransitionTime":"2025-10-03T16:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.362926 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.372233 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.454555 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.454590 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.454598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.454611 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.454620 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:08Z","lastTransitionTime":"2025-10-03T16:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.468824 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb"] Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.469357 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.471357 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.472207 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.482967 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.500586 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.527334 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.542419 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.555507 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.556920 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.556948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.556957 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.556973 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.556984 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:08Z","lastTransitionTime":"2025-10-03T16:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.570765 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.574355 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45hkm\" (UniqueName: \"kubernetes.io/projected/dc860549-945d-4922-a258-99170f6912c1-kube-api-access-45hkm\") pod \"ovnkube-control-plane-749d76644c-425zb\" (UID: \"dc860549-945d-4922-a258-99170f6912c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.574426 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dc860549-945d-4922-a258-99170f6912c1-env-overrides\") pod \"ovnkube-control-plane-749d76644c-425zb\" (UID: \"dc860549-945d-4922-a258-99170f6912c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.574596 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dc860549-945d-4922-a258-99170f6912c1-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-425zb\" (UID: \"dc860549-945d-4922-a258-99170f6912c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.574701 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dc860549-945d-4922-a258-99170f6912c1-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-425zb\" (UID: \"dc860549-945d-4922-a258-99170f6912c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.584101 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.602262 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5b67e587aaf68f1c700802f2ee729e3b8f4bb791de08992d83a5afa0cab2f7c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:06Z\\\",\\\"message\\\":\\\"ers/factory.go:160\\\\nI1003 16:25:06.385860 6031 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:06.386051 6031 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:06.386216 6031 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:06.386312 6031 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 16:25:06.386395 6031 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1003 16:25:06.386431 6031 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1003 16:25:06.386459 6031 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1003 16:25:06.386512 6031 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 16:25:06.386564 6031 factory.go:656] Stopping watch factory\\\\nI1003 16:25:06.386586 6031 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1003 16:25:06.386649 6031 ovnkube.go:599] Stopped ovnkube\\\\nI1003 16:25:06.386466 6031 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 16:25:06.386470 6031 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1003 16:25:06.386522 6031 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:07Z\\\",\\\"message\\\":\\\"0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:07.869637 6152 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 16:25:07.870150 6152 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1003 16:25:07.870170 6152 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1003 16:25:07.870193 6152 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 16:25:07.870214 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1003 16:25:07.870222 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1003 16:25:07.870247 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 16:25:07.870265 6152 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1003 16:25:07.870276 6152 factory.go:656] Stopping watch factory\\\\nI1003 16:25:07.870279 6152 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 16:25:07.870278 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 16:25:07.870289 6152 ovnkube.go:599] Stopped ovnkube\\\\nI1003 16:25:07.870288 6152 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 16:25:07.870294 6152 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.615333 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.627179 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.639810 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.650195 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.660012 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.660070 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.660083 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.660102 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.660115 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:08Z","lastTransitionTime":"2025-10-03T16:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.664914 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.675803 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45hkm\" (UniqueName: \"kubernetes.io/projected/dc860549-945d-4922-a258-99170f6912c1-kube-api-access-45hkm\") pod \"ovnkube-control-plane-749d76644c-425zb\" (UID: \"dc860549-945d-4922-a258-99170f6912c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.675852 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dc860549-945d-4922-a258-99170f6912c1-env-overrides\") pod \"ovnkube-control-plane-749d76644c-425zb\" (UID: \"dc860549-945d-4922-a258-99170f6912c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.675896 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dc860549-945d-4922-a258-99170f6912c1-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-425zb\" (UID: \"dc860549-945d-4922-a258-99170f6912c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.675931 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dc860549-945d-4922-a258-99170f6912c1-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-425zb\" (UID: \"dc860549-945d-4922-a258-99170f6912c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.676719 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dc860549-945d-4922-a258-99170f6912c1-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-425zb\" (UID: \"dc860549-945d-4922-a258-99170f6912c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.676750 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dc860549-945d-4922-a258-99170f6912c1-env-overrides\") pod \"ovnkube-control-plane-749d76644c-425zb\" (UID: \"dc860549-945d-4922-a258-99170f6912c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.680843 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.682473 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dc860549-945d-4922-a258-99170f6912c1-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-425zb\" (UID: \"dc860549-945d-4922-a258-99170f6912c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.694348 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.695401 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45hkm\" (UniqueName: \"kubernetes.io/projected/dc860549-945d-4922-a258-99170f6912c1-kube-api-access-45hkm\") pod \"ovnkube-control-plane-749d76644c-425zb\" (UID: \"dc860549-945d-4922-a258-99170f6912c1\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.707341 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:08Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.763115 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.763207 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.763220 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.763238 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.763248 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:08Z","lastTransitionTime":"2025-10-03T16:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.783289 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" Oct 03 16:25:08 crc kubenswrapper[4744]: W1003 16:25:08.799773 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc860549_945d_4922_a258_99170f6912c1.slice/crio-064e06c7480299f8cafd0750b95a5eea83e75ceac59dae32d47809a5ce5df38d WatchSource:0}: Error finding container 064e06c7480299f8cafd0750b95a5eea83e75ceac59dae32d47809a5ce5df38d: Status 404 returned error can't find the container with id 064e06c7480299f8cafd0750b95a5eea83e75ceac59dae32d47809a5ce5df38d Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.866730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.866775 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.866783 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.866798 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.866808 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:08Z","lastTransitionTime":"2025-10-03T16:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.969684 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.969755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.969766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.969787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:08 crc kubenswrapper[4744]: I1003 16:25:08.969799 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:08Z","lastTransitionTime":"2025-10-03T16:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.072974 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.073046 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.073068 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.073100 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.073120 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:09Z","lastTransitionTime":"2025-10-03T16:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.159173 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovnkube-controller/1.log" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.163804 4744 scope.go:117] "RemoveContainer" containerID="9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.163932 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" event={"ID":"dc860549-945d-4922-a258-99170f6912c1","Type":"ContainerStarted","Data":"064e06c7480299f8cafd0750b95a5eea83e75ceac59dae32d47809a5ce5df38d"} Oct 03 16:25:09 crc kubenswrapper[4744]: E1003 16:25:09.164116 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.176605 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.176690 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.176722 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.176760 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.176780 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:09Z","lastTransitionTime":"2025-10-03T16:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.183146 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.205228 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.225742 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.250810 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.275527 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.279639 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.279683 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.279736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.279757 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.279767 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:09Z","lastTransitionTime":"2025-10-03T16:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.292121 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.308648 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.323031 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.337597 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.360394 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:07Z\\\",\\\"message\\\":\\\"0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:07.869637 6152 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 16:25:07.870150 6152 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1003 16:25:07.870170 6152 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1003 16:25:07.870193 6152 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 16:25:07.870214 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1003 16:25:07.870222 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1003 16:25:07.870247 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 16:25:07.870265 6152 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1003 16:25:07.870276 6152 factory.go:656] Stopping watch factory\\\\nI1003 16:25:07.870279 6152 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 16:25:07.870278 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 16:25:07.870289 6152 ovnkube.go:599] Stopped ovnkube\\\\nI1003 16:25:07.870288 6152 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 16:25:07.870294 6152 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.373830 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.382923 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.382959 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.382969 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.382989 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.383000 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:09Z","lastTransitionTime":"2025-10-03T16:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.388631 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.400173 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.413038 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.427265 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.441140 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.485970 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.486010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.486020 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.486035 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.486045 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:09Z","lastTransitionTime":"2025-10-03T16:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.588247 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.588301 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.588314 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.588331 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.588344 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:09Z","lastTransitionTime":"2025-10-03T16:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.690666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.691295 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.691318 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.691349 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.691372 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:09Z","lastTransitionTime":"2025-10-03T16:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.794386 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.794431 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.794440 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.794455 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.794467 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:09Z","lastTransitionTime":"2025-10-03T16:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.891690 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.891723 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.891803 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:09 crc kubenswrapper[4744]: E1003 16:25:09.891840 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:09 crc kubenswrapper[4744]: E1003 16:25:09.891937 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:09 crc kubenswrapper[4744]: E1003 16:25:09.892290 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.897997 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.898072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.898087 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.898157 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.898181 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:09Z","lastTransitionTime":"2025-10-03T16:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.964728 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-4f7dk"] Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.965256 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:09 crc kubenswrapper[4744]: E1003 16:25:09.965329 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.980159 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.990886 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs\") pod \"network-metrics-daemon-4f7dk\" (UID: \"1008bf15-63d7-45f4-b42b-a4680b7dc232\") " pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:09 crc kubenswrapper[4744]: I1003 16:25:09.990959 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkz2k\" (UniqueName: \"kubernetes.io/projected/1008bf15-63d7-45f4-b42b-a4680b7dc232-kube-api-access-zkz2k\") pod \"network-metrics-daemon-4f7dk\" (UID: \"1008bf15-63d7-45f4-b42b-a4680b7dc232\") " pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.000709 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.000750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.000758 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.000771 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.000781 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:10Z","lastTransitionTime":"2025-10-03T16:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.003223 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:09Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.020461 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.036093 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.061034 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.081226 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.092198 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs\") pod \"network-metrics-daemon-4f7dk\" (UID: \"1008bf15-63d7-45f4-b42b-a4680b7dc232\") " pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.092274 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkz2k\" (UniqueName: \"kubernetes.io/projected/1008bf15-63d7-45f4-b42b-a4680b7dc232-kube-api-access-zkz2k\") pod \"network-metrics-daemon-4f7dk\" (UID: \"1008bf15-63d7-45f4-b42b-a4680b7dc232\") " pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.092385 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.092455 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs podName:1008bf15-63d7-45f4-b42b-a4680b7dc232 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:10.592433552 +0000 UTC m=+36.872309438 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs") pod "network-metrics-daemon-4f7dk" (UID: "1008bf15-63d7-45f4-b42b-a4680b7dc232") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.104269 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.104309 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.104318 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.104336 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.104348 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:10Z","lastTransitionTime":"2025-10-03T16:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.104730 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:07Z\\\",\\\"message\\\":\\\"0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:07.869637 6152 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 16:25:07.870150 6152 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1003 16:25:07.870170 6152 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1003 16:25:07.870193 6152 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 16:25:07.870214 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1003 16:25:07.870222 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1003 16:25:07.870247 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 16:25:07.870265 6152 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1003 16:25:07.870276 6152 factory.go:656] Stopping watch factory\\\\nI1003 16:25:07.870279 6152 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 16:25:07.870278 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 16:25:07.870289 6152 ovnkube.go:599] Stopped ovnkube\\\\nI1003 16:25:07.870288 6152 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 16:25:07.870294 6152 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.112828 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkz2k\" (UniqueName: \"kubernetes.io/projected/1008bf15-63d7-45f4-b42b-a4680b7dc232-kube-api-access-zkz2k\") pod \"network-metrics-daemon-4f7dk\" (UID: \"1008bf15-63d7-45f4-b42b-a4680b7dc232\") " pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.121159 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.134427 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.150732 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.164017 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.171012 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" event={"ID":"dc860549-945d-4922-a258-99170f6912c1","Type":"ContainerStarted","Data":"2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa"} Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.171087 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" event={"ID":"dc860549-945d-4922-a258-99170f6912c1","Type":"ContainerStarted","Data":"9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963"} Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.180119 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.193569 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.206585 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.206637 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.206648 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.206702 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.206713 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:10Z","lastTransitionTime":"2025-10-03T16:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.209180 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.222806 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.234582 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.249028 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.268979 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.284118 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.297640 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.309797 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.309871 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.309885 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.309912 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.309928 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:10Z","lastTransitionTime":"2025-10-03T16:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.315700 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.328097 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.341705 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.355407 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.373549 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.388688 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.402182 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.412228 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.412260 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.412269 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.412285 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.412296 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:10Z","lastTransitionTime":"2025-10-03T16:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.421062 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:07Z\\\",\\\"message\\\":\\\"0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:07.869637 6152 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 16:25:07.870150 6152 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1003 16:25:07.870170 6152 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1003 16:25:07.870193 6152 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 16:25:07.870214 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1003 16:25:07.870222 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1003 16:25:07.870247 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 16:25:07.870265 6152 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1003 16:25:07.870276 6152 factory.go:656] Stopping watch factory\\\\nI1003 16:25:07.870279 6152 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 16:25:07.870278 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 16:25:07.870289 6152 ovnkube.go:599] Stopped ovnkube\\\\nI1003 16:25:07.870288 6152 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 16:25:07.870294 6152 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.435185 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.448020 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.460145 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.470369 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.480668 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.498151 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.498289 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.498332 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.498423 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.498489 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:26.498468313 +0000 UTC m=+52.778344229 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.498548 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:25:26.498529904 +0000 UTC m=+52.778405810 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.498659 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.498706 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:26.498691008 +0000 UTC m=+52.778566904 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.501715 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:10Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.515364 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.515402 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.515412 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.515430 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.515442 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:10Z","lastTransitionTime":"2025-10-03T16:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.599564 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.599614 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.599664 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs\") pod \"network-metrics-daemon-4f7dk\" (UID: \"1008bf15-63d7-45f4-b42b-a4680b7dc232\") " pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.599726 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.599740 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.599754 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.599768 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.599796 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs podName:1008bf15-63d7-45f4-b42b-a4680b7dc232 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:11.599779228 +0000 UTC m=+37.879655124 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs") pod "network-metrics-daemon-4f7dk" (UID: "1008bf15-63d7-45f4-b42b-a4680b7dc232") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.599810 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:26.599804929 +0000 UTC m=+52.879680825 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.599875 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.599939 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.599956 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:10 crc kubenswrapper[4744]: E1003 16:25:10.600077 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:26.600012464 +0000 UTC m=+52.879888440 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.617899 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.617943 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.617954 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.617968 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.617978 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:10Z","lastTransitionTime":"2025-10-03T16:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.720906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.720950 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.720961 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.720978 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.720990 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:10Z","lastTransitionTime":"2025-10-03T16:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.823598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.823630 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.823637 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.823653 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.823663 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:10Z","lastTransitionTime":"2025-10-03T16:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.926570 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.926618 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.926627 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.926642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:10 crc kubenswrapper[4744]: I1003 16:25:10.926651 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:10Z","lastTransitionTime":"2025-10-03T16:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.028590 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.028656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.028669 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.028685 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.028696 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:11Z","lastTransitionTime":"2025-10-03T16:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.131188 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.131253 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.131265 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.131285 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.131298 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:11Z","lastTransitionTime":"2025-10-03T16:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.234020 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.234078 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.234091 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.234112 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.234125 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:11Z","lastTransitionTime":"2025-10-03T16:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.336769 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.336830 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.336841 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.336858 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.336870 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:11Z","lastTransitionTime":"2025-10-03T16:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.439764 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.439807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.439818 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.439834 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.439844 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:11Z","lastTransitionTime":"2025-10-03T16:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.541973 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.542034 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.542045 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.542065 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.542077 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:11Z","lastTransitionTime":"2025-10-03T16:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.582968 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.583014 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.583025 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.583042 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.583052 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:11Z","lastTransitionTime":"2025-10-03T16:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:11 crc kubenswrapper[4744]: E1003 16:25:11.598281 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:11Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.602190 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.602238 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.602248 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.602262 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.602271 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:11Z","lastTransitionTime":"2025-10-03T16:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.611020 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs\") pod \"network-metrics-daemon-4f7dk\" (UID: \"1008bf15-63d7-45f4-b42b-a4680b7dc232\") " pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:11 crc kubenswrapper[4744]: E1003 16:25:11.611171 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:25:11 crc kubenswrapper[4744]: E1003 16:25:11.611224 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs podName:1008bf15-63d7-45f4-b42b-a4680b7dc232 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:13.611207561 +0000 UTC m=+39.891083457 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs") pod "network-metrics-daemon-4f7dk" (UID: "1008bf15-63d7-45f4-b42b-a4680b7dc232") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:25:11 crc kubenswrapper[4744]: E1003 16:25:11.616677 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:11Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.620901 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.620964 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.620977 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.620997 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.621013 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:11Z","lastTransitionTime":"2025-10-03T16:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:11 crc kubenswrapper[4744]: E1003 16:25:11.635629 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:11Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.641435 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.641532 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.641545 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.641566 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.641579 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:11Z","lastTransitionTime":"2025-10-03T16:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:11 crc kubenswrapper[4744]: E1003 16:25:11.659270 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:11Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.664469 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.664555 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.664570 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.664591 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.664636 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:11Z","lastTransitionTime":"2025-10-03T16:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:11 crc kubenswrapper[4744]: E1003 16:25:11.683063 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:11Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:11 crc kubenswrapper[4744]: E1003 16:25:11.683193 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.685294 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.685330 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.685621 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.685948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.685962 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:11Z","lastTransitionTime":"2025-10-03T16:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.790280 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.790326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.790337 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.790354 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.790363 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:11Z","lastTransitionTime":"2025-10-03T16:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.890931 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.890929 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:11 crc kubenswrapper[4744]: E1003 16:25:11.891149 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.891010 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.890980 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:11 crc kubenswrapper[4744]: E1003 16:25:11.891317 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:11 crc kubenswrapper[4744]: E1003 16:25:11.891424 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:11 crc kubenswrapper[4744]: E1003 16:25:11.891558 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.893711 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.893765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.893778 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.893796 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.893809 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:11Z","lastTransitionTime":"2025-10-03T16:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.996861 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.996897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.996908 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.996926 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:11 crc kubenswrapper[4744]: I1003 16:25:11.996937 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:11Z","lastTransitionTime":"2025-10-03T16:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.100005 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.100061 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.100071 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.100092 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.100106 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:12Z","lastTransitionTime":"2025-10-03T16:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.203056 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.203123 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.203134 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.203149 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.203158 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:12Z","lastTransitionTime":"2025-10-03T16:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.305699 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.305742 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.305751 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.305766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.305777 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:12Z","lastTransitionTime":"2025-10-03T16:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.408903 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.408955 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.408968 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.408985 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.408995 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:12Z","lastTransitionTime":"2025-10-03T16:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.511561 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.511617 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.511633 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.511653 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.511664 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:12Z","lastTransitionTime":"2025-10-03T16:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.614031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.614077 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.614087 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.614100 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.614110 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:12Z","lastTransitionTime":"2025-10-03T16:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.716881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.716956 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.716969 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.716987 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.716997 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:12Z","lastTransitionTime":"2025-10-03T16:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.819824 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.819906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.819917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.819936 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.819950 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:12Z","lastTransitionTime":"2025-10-03T16:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.922721 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.922762 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.922771 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.922788 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:12 crc kubenswrapper[4744]: I1003 16:25:12.922797 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:12Z","lastTransitionTime":"2025-10-03T16:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.025588 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.025634 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.025646 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.025670 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.025680 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:13Z","lastTransitionTime":"2025-10-03T16:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.129103 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.129179 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.129198 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.129223 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.129241 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:13Z","lastTransitionTime":"2025-10-03T16:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.231942 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.231996 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.232010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.232027 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.232039 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:13Z","lastTransitionTime":"2025-10-03T16:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.334684 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.334725 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.334740 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.334757 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.334766 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:13Z","lastTransitionTime":"2025-10-03T16:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.437175 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.437233 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.437248 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.437270 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.437287 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:13Z","lastTransitionTime":"2025-10-03T16:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.539807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.539880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.539897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.539914 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.539926 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:13Z","lastTransitionTime":"2025-10-03T16:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.632286 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs\") pod \"network-metrics-daemon-4f7dk\" (UID: \"1008bf15-63d7-45f4-b42b-a4680b7dc232\") " pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:13 crc kubenswrapper[4744]: E1003 16:25:13.632467 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:25:13 crc kubenswrapper[4744]: E1003 16:25:13.632595 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs podName:1008bf15-63d7-45f4-b42b-a4680b7dc232 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:17.63256918 +0000 UTC m=+43.912445076 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs") pod "network-metrics-daemon-4f7dk" (UID: "1008bf15-63d7-45f4-b42b-a4680b7dc232") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.642275 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.642326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.642339 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.642355 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.642367 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:13Z","lastTransitionTime":"2025-10-03T16:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.745257 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.745326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.745338 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.745356 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.745365 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:13Z","lastTransitionTime":"2025-10-03T16:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.847914 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.847994 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.848008 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.848027 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.848048 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:13Z","lastTransitionTime":"2025-10-03T16:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.891579 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.891611 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.891677 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:13 crc kubenswrapper[4744]: E1003 16:25:13.891726 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.891677 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:13 crc kubenswrapper[4744]: E1003 16:25:13.891799 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:13 crc kubenswrapper[4744]: E1003 16:25:13.891933 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:13 crc kubenswrapper[4744]: E1003 16:25:13.892035 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.950341 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.950407 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.950419 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.950435 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:13 crc kubenswrapper[4744]: I1003 16:25:13.950447 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:13Z","lastTransitionTime":"2025-10-03T16:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.053086 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.053140 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.053151 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.053168 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.053177 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:14Z","lastTransitionTime":"2025-10-03T16:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.155283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.155329 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.155340 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.155357 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.155367 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:14Z","lastTransitionTime":"2025-10-03T16:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.257637 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.258353 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.258417 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.258442 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.258454 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:14Z","lastTransitionTime":"2025-10-03T16:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.361158 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.361230 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.361244 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.361262 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.361276 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:14Z","lastTransitionTime":"2025-10-03T16:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.463756 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.463797 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.463806 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.463821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.463829 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:14Z","lastTransitionTime":"2025-10-03T16:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.566953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.567001 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.567018 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.567043 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.567061 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:14Z","lastTransitionTime":"2025-10-03T16:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.670963 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.671021 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.671031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.671049 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.671059 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:14Z","lastTransitionTime":"2025-10-03T16:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.773892 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.773944 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.773955 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.773969 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.773979 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:14Z","lastTransitionTime":"2025-10-03T16:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.876337 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.876423 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.876435 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.876450 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.876459 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:14Z","lastTransitionTime":"2025-10-03T16:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.906072 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:14Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.927394 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:14Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.955670 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:14Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.970957 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:14Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.979540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.979593 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.979602 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.979618 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.979627 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:14Z","lastTransitionTime":"2025-10-03T16:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:14 crc kubenswrapper[4744]: I1003 16:25:14.995424 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:14Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.009918 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.024146 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.038778 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.050836 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.069302 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:07Z\\\",\\\"message\\\":\\\"0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:07.869637 6152 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 16:25:07.870150 6152 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1003 16:25:07.870170 6152 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1003 16:25:07.870193 6152 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 16:25:07.870214 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1003 16:25:07.870222 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1003 16:25:07.870247 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 16:25:07.870265 6152 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1003 16:25:07.870276 6152 factory.go:656] Stopping watch factory\\\\nI1003 16:25:07.870279 6152 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 16:25:07.870278 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 16:25:07.870289 6152 ovnkube.go:599] Stopped ovnkube\\\\nI1003 16:25:07.870288 6152 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 16:25:07.870294 6152 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.081376 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.081428 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.081436 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.081451 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.081459 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:15Z","lastTransitionTime":"2025-10-03T16:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.084912 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.099601 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.115167 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.126453 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.139767 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.152596 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.164828 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.183864 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.183906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.183934 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.183965 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.183976 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:15Z","lastTransitionTime":"2025-10-03T16:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.286401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.286443 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.286454 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.286470 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.286481 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:15Z","lastTransitionTime":"2025-10-03T16:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.389963 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.390017 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.390030 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.390057 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.390071 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:15Z","lastTransitionTime":"2025-10-03T16:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.492461 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.492584 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.492603 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.492634 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.492655 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:15Z","lastTransitionTime":"2025-10-03T16:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.596041 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.596093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.596104 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.596123 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.596137 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:15Z","lastTransitionTime":"2025-10-03T16:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.699255 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.699337 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.699356 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.699388 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.699409 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:15Z","lastTransitionTime":"2025-10-03T16:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.803044 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.803131 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.803158 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.803195 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.803222 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:15Z","lastTransitionTime":"2025-10-03T16:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.891481 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.891591 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.891602 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:15 crc kubenswrapper[4744]: E1003 16:25:15.891749 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.891810 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:15 crc kubenswrapper[4744]: E1003 16:25:15.891969 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:15 crc kubenswrapper[4744]: E1003 16:25:15.892102 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:15 crc kubenswrapper[4744]: E1003 16:25:15.892210 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.906905 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.906955 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.906966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.907017 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:15 crc kubenswrapper[4744]: I1003 16:25:15.907032 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:15Z","lastTransitionTime":"2025-10-03T16:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.011210 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.011282 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.011293 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.011311 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.011321 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:16Z","lastTransitionTime":"2025-10-03T16:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.114810 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.114870 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.114883 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.114902 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.114913 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:16Z","lastTransitionTime":"2025-10-03T16:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.217425 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.217484 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.217510 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.217526 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.217539 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:16Z","lastTransitionTime":"2025-10-03T16:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.319696 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.319748 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.319757 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.319773 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.319783 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:16Z","lastTransitionTime":"2025-10-03T16:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.422399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.422463 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.422477 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.422518 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.422529 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:16Z","lastTransitionTime":"2025-10-03T16:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.524827 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.524865 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.524876 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.524894 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.524905 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:16Z","lastTransitionTime":"2025-10-03T16:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.568066 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.568777 4744 scope.go:117] "RemoveContainer" containerID="9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb" Oct 03 16:25:16 crc kubenswrapper[4744]: E1003 16:25:16.569008 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.627641 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.627705 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.627723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.627746 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.627766 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:16Z","lastTransitionTime":"2025-10-03T16:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.730490 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.730560 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.730569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.730583 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.730592 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:16Z","lastTransitionTime":"2025-10-03T16:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.833034 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.833080 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.833092 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.833108 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.833120 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:16Z","lastTransitionTime":"2025-10-03T16:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.935202 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.935251 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.935260 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.935277 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:16 crc kubenswrapper[4744]: I1003 16:25:16.935287 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:16Z","lastTransitionTime":"2025-10-03T16:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.037787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.037844 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.037853 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.037867 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.037877 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:17Z","lastTransitionTime":"2025-10-03T16:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.140512 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.140552 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.140560 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.140576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.140584 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:17Z","lastTransitionTime":"2025-10-03T16:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.243473 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.243529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.243537 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.243551 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.243561 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:17Z","lastTransitionTime":"2025-10-03T16:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.346487 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.346557 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.346565 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.346597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.346608 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:17Z","lastTransitionTime":"2025-10-03T16:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.449362 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.449424 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.449433 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.449447 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.449454 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:17Z","lastTransitionTime":"2025-10-03T16:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.552121 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.552179 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.552191 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.552210 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.552222 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:17Z","lastTransitionTime":"2025-10-03T16:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.654819 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.654868 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.654879 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.654896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.654910 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:17Z","lastTransitionTime":"2025-10-03T16:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.680664 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs\") pod \"network-metrics-daemon-4f7dk\" (UID: \"1008bf15-63d7-45f4-b42b-a4680b7dc232\") " pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:17 crc kubenswrapper[4744]: E1003 16:25:17.680790 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:25:17 crc kubenswrapper[4744]: E1003 16:25:17.680852 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs podName:1008bf15-63d7-45f4-b42b-a4680b7dc232 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:25.680833873 +0000 UTC m=+51.960709769 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs") pod "network-metrics-daemon-4f7dk" (UID: "1008bf15-63d7-45f4-b42b-a4680b7dc232") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.757176 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.757223 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.757232 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.757253 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.757268 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:17Z","lastTransitionTime":"2025-10-03T16:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.859529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.859581 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.859591 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.859608 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.859619 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:17Z","lastTransitionTime":"2025-10-03T16:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.891300 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.891359 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:17 crc kubenswrapper[4744]: E1003 16:25:17.891427 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.891303 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:17 crc kubenswrapper[4744]: E1003 16:25:17.891516 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.891361 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:17 crc kubenswrapper[4744]: E1003 16:25:17.891615 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:17 crc kubenswrapper[4744]: E1003 16:25:17.891805 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.961813 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.961880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.961889 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.961903 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:17 crc kubenswrapper[4744]: I1003 16:25:17.961912 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:17Z","lastTransitionTime":"2025-10-03T16:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.064401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.064450 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.064460 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.064475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.064484 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:18Z","lastTransitionTime":"2025-10-03T16:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.166343 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.166392 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.166409 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.166429 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.166439 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:18Z","lastTransitionTime":"2025-10-03T16:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.268947 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.268994 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.269004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.269020 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.269032 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:18Z","lastTransitionTime":"2025-10-03T16:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.371581 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.371984 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.371998 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.372015 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.372028 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:18Z","lastTransitionTime":"2025-10-03T16:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.474074 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.474117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.474127 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.474146 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.474157 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:18Z","lastTransitionTime":"2025-10-03T16:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.576396 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.576433 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.576441 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.576457 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.576466 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:18Z","lastTransitionTime":"2025-10-03T16:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.678603 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.678706 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.678720 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.678740 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.678756 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:18Z","lastTransitionTime":"2025-10-03T16:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.781326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.781372 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.781381 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.781399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.781409 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:18Z","lastTransitionTime":"2025-10-03T16:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.883395 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.883463 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.883483 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.883525 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.883543 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:18Z","lastTransitionTime":"2025-10-03T16:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.986443 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.986487 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.986522 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.986542 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:18 crc kubenswrapper[4744]: I1003 16:25:18.986555 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:18Z","lastTransitionTime":"2025-10-03T16:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.089822 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.089874 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.089883 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.089902 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.089911 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:19Z","lastTransitionTime":"2025-10-03T16:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.192245 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.192308 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.192318 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.192337 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.192350 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:19Z","lastTransitionTime":"2025-10-03T16:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.296591 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.296666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.296688 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.296716 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.296737 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:19Z","lastTransitionTime":"2025-10-03T16:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.399671 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.399730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.399748 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.399771 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.399790 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:19Z","lastTransitionTime":"2025-10-03T16:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.502435 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.502513 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.502529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.502559 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.502576 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:19Z","lastTransitionTime":"2025-10-03T16:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.606328 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.606454 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.606584 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.606675 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.606732 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:19Z","lastTransitionTime":"2025-10-03T16:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.710480 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.710604 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.710630 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.710667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.710692 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:19Z","lastTransitionTime":"2025-10-03T16:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.814772 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.814815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.814823 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.814839 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.814851 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:19Z","lastTransitionTime":"2025-10-03T16:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.890932 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.890973 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.891012 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:19 crc kubenswrapper[4744]: E1003 16:25:19.891154 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.891198 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:19 crc kubenswrapper[4744]: E1003 16:25:19.891343 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:19 crc kubenswrapper[4744]: E1003 16:25:19.891461 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:19 crc kubenswrapper[4744]: E1003 16:25:19.891604 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.918098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.918146 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.918156 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.918175 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:19 crc kubenswrapper[4744]: I1003 16:25:19.918186 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:19Z","lastTransitionTime":"2025-10-03T16:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.021376 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.021453 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.021465 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.021483 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.021512 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:20Z","lastTransitionTime":"2025-10-03T16:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.124422 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.124480 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.124491 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.124527 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.124538 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:20Z","lastTransitionTime":"2025-10-03T16:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.226953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.226991 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.227002 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.227018 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.227027 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:20Z","lastTransitionTime":"2025-10-03T16:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.330827 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.330875 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.330884 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.330900 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.330911 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:20Z","lastTransitionTime":"2025-10-03T16:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.433609 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.433657 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.433669 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.433686 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.433698 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:20Z","lastTransitionTime":"2025-10-03T16:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.536423 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.536475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.536485 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.536524 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.536536 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:20Z","lastTransitionTime":"2025-10-03T16:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.638864 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.638934 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.638947 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.638975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.638994 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:20Z","lastTransitionTime":"2025-10-03T16:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.741253 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.741299 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.741309 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.741327 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.741336 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:20Z","lastTransitionTime":"2025-10-03T16:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.843916 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.843974 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.843986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.844022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.844038 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:20Z","lastTransitionTime":"2025-10-03T16:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.946242 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.946273 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.946283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.946295 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:20 crc kubenswrapper[4744]: I1003 16:25:20.946304 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:20Z","lastTransitionTime":"2025-10-03T16:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.049144 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.049200 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.049211 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.049232 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.049244 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:21Z","lastTransitionTime":"2025-10-03T16:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.152727 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.152862 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.152876 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.152894 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.152906 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:21Z","lastTransitionTime":"2025-10-03T16:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.255963 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.256058 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.256093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.256127 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.256154 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:21Z","lastTransitionTime":"2025-10-03T16:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.359311 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.359362 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.359374 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.359391 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.359402 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:21Z","lastTransitionTime":"2025-10-03T16:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.462821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.462865 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.462874 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.462889 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.462902 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:21Z","lastTransitionTime":"2025-10-03T16:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.565578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.565668 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.565683 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.565701 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.565713 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:21Z","lastTransitionTime":"2025-10-03T16:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.668739 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.668838 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.668848 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.668866 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.668878 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:21Z","lastTransitionTime":"2025-10-03T16:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.772165 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.772235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.772252 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.772274 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.772289 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:21Z","lastTransitionTime":"2025-10-03T16:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.876344 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.876401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.876413 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.876433 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.876445 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:21Z","lastTransitionTime":"2025-10-03T16:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.891765 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.891827 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.891881 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.891779 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:21 crc kubenswrapper[4744]: E1003 16:25:21.891969 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:21 crc kubenswrapper[4744]: E1003 16:25:21.892072 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:21 crc kubenswrapper[4744]: E1003 16:25:21.892135 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:21 crc kubenswrapper[4744]: E1003 16:25:21.892176 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.940692 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.940765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.940775 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.940792 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.940801 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:21Z","lastTransitionTime":"2025-10-03T16:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:21 crc kubenswrapper[4744]: E1003 16:25:21.954558 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:21Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.958636 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.958666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.958675 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.958691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.958701 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:21Z","lastTransitionTime":"2025-10-03T16:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:21 crc kubenswrapper[4744]: E1003 16:25:21.971661 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:21Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.975530 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.975562 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.975571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.975586 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.975598 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:21Z","lastTransitionTime":"2025-10-03T16:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:21 crc kubenswrapper[4744]: E1003 16:25:21.986603 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:21Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.993029 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.993144 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.993237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.993274 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:21 crc kubenswrapper[4744]: I1003 16:25:21.993324 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:21Z","lastTransitionTime":"2025-10-03T16:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:22 crc kubenswrapper[4744]: E1003 16:25:22.008319 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:22Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.012085 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.012117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.012125 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.012140 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.012167 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:22Z","lastTransitionTime":"2025-10-03T16:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:22 crc kubenswrapper[4744]: E1003 16:25:22.029286 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:22Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:22 crc kubenswrapper[4744]: E1003 16:25:22.029481 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.032576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.032620 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.032633 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.032656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.032672 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:22Z","lastTransitionTime":"2025-10-03T16:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.135546 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.135602 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.135614 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.135636 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.135650 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:22Z","lastTransitionTime":"2025-10-03T16:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.238942 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.238992 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.239001 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.239019 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.239029 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:22Z","lastTransitionTime":"2025-10-03T16:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.342053 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.342102 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.342115 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.342138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.342153 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:22Z","lastTransitionTime":"2025-10-03T16:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.445187 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.445251 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.445263 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.445285 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.445301 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:22Z","lastTransitionTime":"2025-10-03T16:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.548092 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.548154 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.548167 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.548192 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.548205 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:22Z","lastTransitionTime":"2025-10-03T16:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.652352 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.652430 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.652450 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.652482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.652524 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:22Z","lastTransitionTime":"2025-10-03T16:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.755257 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.755336 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.755348 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.755364 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.755372 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:22Z","lastTransitionTime":"2025-10-03T16:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.858522 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.858571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.858580 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.858596 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.858609 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:22Z","lastTransitionTime":"2025-10-03T16:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.961031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.961071 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.961080 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.961094 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:22 crc kubenswrapper[4744]: I1003 16:25:22.961103 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:22Z","lastTransitionTime":"2025-10-03T16:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.063476 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.063537 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.063549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.063565 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.063576 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:23Z","lastTransitionTime":"2025-10-03T16:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.165654 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.165726 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.165739 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.165754 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.165765 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:23Z","lastTransitionTime":"2025-10-03T16:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.267632 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.267713 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.267749 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.267770 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.267781 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:23Z","lastTransitionTime":"2025-10-03T16:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.370883 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.370940 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.370953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.370974 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.370986 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:23Z","lastTransitionTime":"2025-10-03T16:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.473331 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.473376 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.473387 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.473404 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.473416 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:23Z","lastTransitionTime":"2025-10-03T16:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.576469 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.576555 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.576569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.576587 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.576600 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:23Z","lastTransitionTime":"2025-10-03T16:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.679442 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.679487 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.679511 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.679526 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.679536 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:23Z","lastTransitionTime":"2025-10-03T16:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.782409 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.782461 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.782544 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.782569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.782581 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:23Z","lastTransitionTime":"2025-10-03T16:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.854865 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.866139 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.872281 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:23Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.885026 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.885070 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.885084 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.885096 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.885106 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:23Z","lastTransitionTime":"2025-10-03T16:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.891712 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.891742 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.891742 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.891846 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:23 crc kubenswrapper[4744]: E1003 16:25:23.892035 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:23 crc kubenswrapper[4744]: E1003 16:25:23.892130 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:23 crc kubenswrapper[4744]: E1003 16:25:23.892230 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:23 crc kubenswrapper[4744]: E1003 16:25:23.892297 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.894071 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:23Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.906771 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:23Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.932895 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:23Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.953233 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:23Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.971204 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:23Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.988198 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.988258 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.988267 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.988282 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.988291 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:23Z","lastTransitionTime":"2025-10-03T16:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:23 crc kubenswrapper[4744]: I1003 16:25:23.991324 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:07Z\\\",\\\"message\\\":\\\"0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:07.869637 6152 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 16:25:07.870150 6152 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1003 16:25:07.870170 6152 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1003 16:25:07.870193 6152 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 16:25:07.870214 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1003 16:25:07.870222 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1003 16:25:07.870247 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 16:25:07.870265 6152 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1003 16:25:07.870276 6152 factory.go:656] Stopping watch factory\\\\nI1003 16:25:07.870279 6152 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 16:25:07.870278 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 16:25:07.870289 6152 ovnkube.go:599] Stopped ovnkube\\\\nI1003 16:25:07.870288 6152 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 16:25:07.870294 6152 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:23Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.007792 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.022479 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.038124 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.053425 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.065859 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.080642 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.091207 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.091293 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.091311 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.091334 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.091350 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:24Z","lastTransitionTime":"2025-10-03T16:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.095419 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.109086 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.120096 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.131042 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.194691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.194733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.194743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.194761 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.194773 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:24Z","lastTransitionTime":"2025-10-03T16:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.297628 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.297681 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.297693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.297719 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.297733 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:24Z","lastTransitionTime":"2025-10-03T16:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.401101 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.401167 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.401180 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.401197 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.401208 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:24Z","lastTransitionTime":"2025-10-03T16:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.504614 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.504667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.504679 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.504701 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.504719 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:24Z","lastTransitionTime":"2025-10-03T16:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.608074 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.608120 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.608130 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.608149 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.608163 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:24Z","lastTransitionTime":"2025-10-03T16:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.712054 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.712110 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.712121 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.712142 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.712153 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:24Z","lastTransitionTime":"2025-10-03T16:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.815407 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.815483 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.815536 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.815568 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.815590 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:24Z","lastTransitionTime":"2025-10-03T16:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.907691 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.917733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.917803 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.917823 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.917869 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.917892 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:24Z","lastTransitionTime":"2025-10-03T16:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.923244 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.936395 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.951824 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.972866 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:24 crc kubenswrapper[4744]: I1003 16:25:24.989737 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:24Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.003444 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:25Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.017985 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:25Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.021187 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.021265 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.021300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.021326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.021342 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:25Z","lastTransitionTime":"2025-10-03T16:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.042008 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:25Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.056658 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:25Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.091398 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:25Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.109097 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:25Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.125378 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.125443 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.125460 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.125487 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.125538 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:25Z","lastTransitionTime":"2025-10-03T16:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.130229 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:25Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.149370 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:25Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.164377 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:25Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.195886 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:07Z\\\",\\\"message\\\":\\\"0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:07.869637 6152 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 16:25:07.870150 6152 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1003 16:25:07.870170 6152 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1003 16:25:07.870193 6152 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 16:25:07.870214 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1003 16:25:07.870222 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1003 16:25:07.870247 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 16:25:07.870265 6152 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1003 16:25:07.870276 6152 factory.go:656] Stopping watch factory\\\\nI1003 16:25:07.870279 6152 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 16:25:07.870278 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 16:25:07.870289 6152 ovnkube.go:599] Stopped ovnkube\\\\nI1003 16:25:07.870288 6152 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 16:25:07.870294 6152 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:25Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.210319 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81781145-978c-49ef-bb8e-4bf4f5eab778\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011b2e41b97e95ff1a11ff71d90fc00813ab18eac48b6cc7a987039494649a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6b1c34f972682f8afe389a96e2b35b848e4ceb02cb9379fa91c764f2cf11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52cbfb9a1c58ce525029b4226599a8e258453a8c5837643a6fd1efe2441641a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:25Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.227914 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.228065 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.228177 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.228289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.228400 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:25Z","lastTransitionTime":"2025-10-03T16:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.229080 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:25Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.330607 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.330651 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.330662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.330678 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.330689 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:25Z","lastTransitionTime":"2025-10-03T16:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.432566 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.432601 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.432610 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.432623 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.432631 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:25Z","lastTransitionTime":"2025-10-03T16:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.534630 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.534667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.534678 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.534691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.534700 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:25Z","lastTransitionTime":"2025-10-03T16:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.637454 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.637842 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.637935 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.638033 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.638178 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:25Z","lastTransitionTime":"2025-10-03T16:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.740961 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.741530 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.741660 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.741743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.741801 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:25Z","lastTransitionTime":"2025-10-03T16:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.773655 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs\") pod \"network-metrics-daemon-4f7dk\" (UID: \"1008bf15-63d7-45f4-b42b-a4680b7dc232\") " pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:25 crc kubenswrapper[4744]: E1003 16:25:25.773797 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:25:25 crc kubenswrapper[4744]: E1003 16:25:25.773844 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs podName:1008bf15-63d7-45f4-b42b-a4680b7dc232 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:41.77382998 +0000 UTC m=+68.053705876 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs") pod "network-metrics-daemon-4f7dk" (UID: "1008bf15-63d7-45f4-b42b-a4680b7dc232") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.844839 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.844917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.844943 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.844975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.844993 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:25Z","lastTransitionTime":"2025-10-03T16:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.891865 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.891920 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.891933 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.891874 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:25 crc kubenswrapper[4744]: E1003 16:25:25.892081 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:25 crc kubenswrapper[4744]: E1003 16:25:25.892181 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:25 crc kubenswrapper[4744]: E1003 16:25:25.892285 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:25 crc kubenswrapper[4744]: E1003 16:25:25.892377 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.947371 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.947419 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.947429 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.947445 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:25 crc kubenswrapper[4744]: I1003 16:25:25.947456 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:25Z","lastTransitionTime":"2025-10-03T16:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.050972 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.051035 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.051049 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.051067 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.051080 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:26Z","lastTransitionTime":"2025-10-03T16:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.154031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.154094 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.154107 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.154131 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.154147 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:26Z","lastTransitionTime":"2025-10-03T16:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.257701 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.257756 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.257766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.257787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.257798 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:26Z","lastTransitionTime":"2025-10-03T16:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.361343 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.361404 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.361417 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.361439 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.361453 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:26Z","lastTransitionTime":"2025-10-03T16:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.464474 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.464548 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.464563 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.464585 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.464601 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:26Z","lastTransitionTime":"2025-10-03T16:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.568409 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.568555 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.568586 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.568619 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.568642 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:26Z","lastTransitionTime":"2025-10-03T16:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.583259 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.583606 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.583750 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:26 crc kubenswrapper[4744]: E1003 16:25:26.584000 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:25:26 crc kubenswrapper[4744]: E1003 16:25:26.584033 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:25:26 crc kubenswrapper[4744]: E1003 16:25:26.584041 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:25:58.583995045 +0000 UTC m=+84.863870941 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:25:26 crc kubenswrapper[4744]: E1003 16:25:26.584305 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:58.584291192 +0000 UTC m=+84.864167088 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:25:26 crc kubenswrapper[4744]: E1003 16:25:26.584396 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:58.584384804 +0000 UTC m=+84.864260700 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.671930 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.672365 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.672439 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.672553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.672661 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:26Z","lastTransitionTime":"2025-10-03T16:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.684782 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.684850 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:26 crc kubenswrapper[4744]: E1003 16:25:26.685189 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:25:26 crc kubenswrapper[4744]: E1003 16:25:26.685240 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:25:26 crc kubenswrapper[4744]: E1003 16:25:26.685268 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:26 crc kubenswrapper[4744]: E1003 16:25:26.685412 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:58.685350201 +0000 UTC m=+84.965226137 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:26 crc kubenswrapper[4744]: E1003 16:25:26.685777 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:25:26 crc kubenswrapper[4744]: E1003 16:25:26.685910 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:25:26 crc kubenswrapper[4744]: E1003 16:25:26.685983 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:26 crc kubenswrapper[4744]: E1003 16:25:26.686149 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 16:25:58.68612213 +0000 UTC m=+84.965998226 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.776055 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.776728 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.776769 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.776799 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.776820 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:26Z","lastTransitionTime":"2025-10-03T16:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.880841 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.880925 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.880946 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.880978 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.881001 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:26Z","lastTransitionTime":"2025-10-03T16:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.984303 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.984368 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.984381 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.984401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:26 crc kubenswrapper[4744]: I1003 16:25:26.984414 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:26Z","lastTransitionTime":"2025-10-03T16:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.087542 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.087574 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.087582 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.087599 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.087611 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:27Z","lastTransitionTime":"2025-10-03T16:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.191344 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.191408 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.191427 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.191454 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.191475 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:27Z","lastTransitionTime":"2025-10-03T16:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.294911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.294969 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.294979 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.295014 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.295029 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:27Z","lastTransitionTime":"2025-10-03T16:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.397604 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.397673 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.397691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.397724 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.397744 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:27Z","lastTransitionTime":"2025-10-03T16:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.500693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.500764 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.500789 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.500820 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.500843 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:27Z","lastTransitionTime":"2025-10-03T16:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.603754 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.603826 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.603850 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.603884 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.603907 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:27Z","lastTransitionTime":"2025-10-03T16:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.706945 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.707035 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.707069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.707112 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.707138 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:27Z","lastTransitionTime":"2025-10-03T16:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.810637 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.810721 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.810747 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.810782 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.810813 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:27Z","lastTransitionTime":"2025-10-03T16:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.891182 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.891281 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.891187 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.891178 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:27 crc kubenswrapper[4744]: E1003 16:25:27.891407 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:27 crc kubenswrapper[4744]: E1003 16:25:27.891532 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:27 crc kubenswrapper[4744]: E1003 16:25:27.891999 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:27 crc kubenswrapper[4744]: E1003 16:25:27.892318 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.913604 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.913661 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.913673 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.913694 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:27 crc kubenswrapper[4744]: I1003 16:25:27.913712 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:27Z","lastTransitionTime":"2025-10-03T16:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.017891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.017946 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.017960 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.017982 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.017996 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:28Z","lastTransitionTime":"2025-10-03T16:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.121503 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.121556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.121570 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.121597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.121614 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:28Z","lastTransitionTime":"2025-10-03T16:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.224242 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.224294 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.224305 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.224327 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.224338 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:28Z","lastTransitionTime":"2025-10-03T16:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.327018 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.327054 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.327062 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.327075 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.327085 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:28Z","lastTransitionTime":"2025-10-03T16:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.431320 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.431402 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.431421 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.431458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.431482 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:28Z","lastTransitionTime":"2025-10-03T16:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.535431 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.535514 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.535527 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.535550 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.535563 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:28Z","lastTransitionTime":"2025-10-03T16:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.638113 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.638155 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.638165 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.638183 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.638199 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:28Z","lastTransitionTime":"2025-10-03T16:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.741978 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.742055 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.742078 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.742110 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.742132 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:28Z","lastTransitionTime":"2025-10-03T16:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.845397 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.845478 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.845544 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.845581 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.845607 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:28Z","lastTransitionTime":"2025-10-03T16:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.949048 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.949113 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.949132 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.949160 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:28 crc kubenswrapper[4744]: I1003 16:25:28.949191 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:28Z","lastTransitionTime":"2025-10-03T16:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.052322 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.052412 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.052442 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.052478 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.052546 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:29Z","lastTransitionTime":"2025-10-03T16:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.155552 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.155605 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.155618 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.155639 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.155655 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:29Z","lastTransitionTime":"2025-10-03T16:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.258478 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.258555 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.258565 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.258585 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.258596 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:29Z","lastTransitionTime":"2025-10-03T16:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.361765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.361845 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.361868 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.361900 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.361923 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:29Z","lastTransitionTime":"2025-10-03T16:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.464876 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.464963 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.464983 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.465009 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.465027 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:29Z","lastTransitionTime":"2025-10-03T16:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.567201 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.567267 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.567282 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.567324 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.567337 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:29Z","lastTransitionTime":"2025-10-03T16:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.670352 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.670450 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.670469 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.670546 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.670570 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:29Z","lastTransitionTime":"2025-10-03T16:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.773787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.773850 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.773864 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.773887 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.773902 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:29Z","lastTransitionTime":"2025-10-03T16:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.877681 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.877784 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.877804 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.877821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.877834 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:29Z","lastTransitionTime":"2025-10-03T16:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.891067 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.891117 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.891067 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.891246 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:29 crc kubenswrapper[4744]: E1003 16:25:29.891451 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:29 crc kubenswrapper[4744]: E1003 16:25:29.891662 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:29 crc kubenswrapper[4744]: E1003 16:25:29.891746 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:29 crc kubenswrapper[4744]: E1003 16:25:29.891990 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.981928 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.982009 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.982035 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.982064 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:29 crc kubenswrapper[4744]: I1003 16:25:29.982086 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:29Z","lastTransitionTime":"2025-10-03T16:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.085520 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.085556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.085565 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.085578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.085589 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:30Z","lastTransitionTime":"2025-10-03T16:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.188576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.188645 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.188658 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.188679 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.188693 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:30Z","lastTransitionTime":"2025-10-03T16:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.291395 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.291443 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.291456 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.291476 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.291490 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:30Z","lastTransitionTime":"2025-10-03T16:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.395682 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.395761 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.395782 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.395813 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.395832 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:30Z","lastTransitionTime":"2025-10-03T16:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.499275 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.499360 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.499381 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.499415 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.499437 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:30Z","lastTransitionTime":"2025-10-03T16:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.603339 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.603416 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.603437 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.603470 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.603491 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:30Z","lastTransitionTime":"2025-10-03T16:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.707103 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.707170 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.707188 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.707217 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.707239 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:30Z","lastTransitionTime":"2025-10-03T16:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.811939 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.812065 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.812087 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.812155 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.812178 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:30Z","lastTransitionTime":"2025-10-03T16:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.915182 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.915278 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.915301 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.915335 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:30 crc kubenswrapper[4744]: I1003 16:25:30.915360 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:30Z","lastTransitionTime":"2025-10-03T16:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.019881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.019961 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.019989 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.020023 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.020049 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:31Z","lastTransitionTime":"2025-10-03T16:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.123291 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.123370 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.123383 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.123404 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.123418 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:31Z","lastTransitionTime":"2025-10-03T16:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.227205 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.227265 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.227278 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.227300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.227325 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:31Z","lastTransitionTime":"2025-10-03T16:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.329863 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.329911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.329921 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.329937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.329946 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:31Z","lastTransitionTime":"2025-10-03T16:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.433294 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.433354 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.433372 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.433402 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.433425 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:31Z","lastTransitionTime":"2025-10-03T16:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.536734 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.536820 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.536844 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.536878 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.536900 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:31Z","lastTransitionTime":"2025-10-03T16:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.640488 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.640580 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.640597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.640626 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.640644 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:31Z","lastTransitionTime":"2025-10-03T16:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.743982 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.744036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.744050 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.744069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.744081 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:31Z","lastTransitionTime":"2025-10-03T16:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.846674 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.846741 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.846751 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.846772 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.846785 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:31Z","lastTransitionTime":"2025-10-03T16:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.891451 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.891574 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.891580 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.891765 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:31 crc kubenswrapper[4744]: E1003 16:25:31.891761 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:31 crc kubenswrapper[4744]: E1003 16:25:31.891870 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:31 crc kubenswrapper[4744]: E1003 16:25:31.891986 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.892722 4744 scope.go:117] "RemoveContainer" containerID="9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb" Oct 03 16:25:31 crc kubenswrapper[4744]: E1003 16:25:31.892770 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.949520 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.950040 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.950055 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.950080 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:31 crc kubenswrapper[4744]: I1003 16:25:31.950098 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:31Z","lastTransitionTime":"2025-10-03T16:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.052949 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.053004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.053015 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.053037 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.053056 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:32Z","lastTransitionTime":"2025-10-03T16:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.156011 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.156056 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.156066 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.156083 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.156093 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:32Z","lastTransitionTime":"2025-10-03T16:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.231393 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.231702 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.231841 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.231922 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.231995 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:32Z","lastTransitionTime":"2025-10-03T16:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.246773 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovnkube-controller/1.log" Oct 03 16:25:32 crc kubenswrapper[4744]: E1003 16:25:32.248005 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.250214 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerStarted","Data":"ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc"} Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.251008 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.252696 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.252733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.252747 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.252766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.252782 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:32Z","lastTransitionTime":"2025-10-03T16:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.267884 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: E1003 16:25:32.267967 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.272764 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.272825 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.272841 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.272867 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.272883 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:32Z","lastTransitionTime":"2025-10-03T16:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.288301 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: E1003 16:25:32.288489 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.295062 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.295106 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.295117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.295137 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.295151 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:32Z","lastTransitionTime":"2025-10-03T16:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.306899 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: E1003 16:25:32.309847 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.315403 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.315462 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.315480 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.315531 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.315551 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:32Z","lastTransitionTime":"2025-10-03T16:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.322518 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: E1003 16:25:32.330674 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: E1003 16:25:32.330888 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.333038 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.333217 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.333322 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.333410 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.333513 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:32Z","lastTransitionTime":"2025-10-03T16:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.342804 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.359845 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.388263 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:07Z\\\",\\\"message\\\":\\\"0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:07.869637 6152 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 16:25:07.870150 6152 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1003 16:25:07.870170 6152 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1003 16:25:07.870193 6152 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 16:25:07.870214 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1003 16:25:07.870222 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1003 16:25:07.870247 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 16:25:07.870265 6152 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1003 16:25:07.870276 6152 factory.go:656] Stopping watch factory\\\\nI1003 16:25:07.870279 6152 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 16:25:07.870278 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 16:25:07.870289 6152 ovnkube.go:599] Stopped ovnkube\\\\nI1003 16:25:07.870288 6152 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 16:25:07.870294 6152 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.403095 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81781145-978c-49ef-bb8e-4bf4f5eab778\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011b2e41b97e95ff1a11ff71d90fc00813ab18eac48b6cc7a987039494649a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6b1c34f972682f8afe389a96e2b35b848e4ceb02cb9379fa91c764f2cf11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52cbfb9a1c58ce525029b4226599a8e258453a8c5837643a6fd1efe2441641a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.416879 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.433780 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.435969 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.436008 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.436021 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.436043 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.436056 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:32Z","lastTransitionTime":"2025-10-03T16:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.449645 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.465625 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.479070 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.491024 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.507264 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.523067 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.537366 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.538590 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.538621 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.538630 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.538648 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.538660 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:32Z","lastTransitionTime":"2025-10-03T16:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.551340 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:32Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.642331 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.642388 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.642400 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.642420 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.642436 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:32Z","lastTransitionTime":"2025-10-03T16:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.745879 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.746150 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.746232 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.746320 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.746404 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:32Z","lastTransitionTime":"2025-10-03T16:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.849798 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.849854 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.849868 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.849891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.849907 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:32Z","lastTransitionTime":"2025-10-03T16:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.953291 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.953402 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.953412 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.953431 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:32 crc kubenswrapper[4744]: I1003 16:25:32.953443 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:32Z","lastTransitionTime":"2025-10-03T16:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.057271 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.057357 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.057371 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.057398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.057412 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:33Z","lastTransitionTime":"2025-10-03T16:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.160817 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.160895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.160915 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.160942 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.160961 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:33Z","lastTransitionTime":"2025-10-03T16:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.258451 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovnkube-controller/2.log" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.260206 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovnkube-controller/1.log" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.267569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.267621 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.267637 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.267659 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.267672 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:33Z","lastTransitionTime":"2025-10-03T16:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.271206 4744 generic.go:334] "Generic (PLEG): container finished" podID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerID="ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc" exitCode=1 Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.271286 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerDied","Data":"ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc"} Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.271346 4744 scope.go:117] "RemoveContainer" containerID="9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.273096 4744 scope.go:117] "RemoveContainer" containerID="ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc" Oct 03 16:25:33 crc kubenswrapper[4744]: E1003 16:25:33.273444 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.291434 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.309058 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.329077 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.346794 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.362783 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.371177 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.371223 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.371233 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.371251 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.371262 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:33Z","lastTransitionTime":"2025-10-03T16:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.381922 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.398450 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.419020 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.432845 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.467145 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.474225 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.474300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.474326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.474358 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.474377 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:33Z","lastTransitionTime":"2025-10-03T16:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.486408 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.503732 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.521870 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.535999 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.556890 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b8221246e71865d83a571ac12041eaabce6ba54975c5f3eec0ed1fd4cd10ceb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:07Z\\\",\\\"message\\\":\\\"0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 16:25:07.869637 6152 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 16:25:07.870150 6152 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1003 16:25:07.870170 6152 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1003 16:25:07.870193 6152 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1003 16:25:07.870214 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1003 16:25:07.870222 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1003 16:25:07.870247 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1003 16:25:07.870265 6152 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1003 16:25:07.870276 6152 factory.go:656] Stopping watch factory\\\\nI1003 16:25:07.870279 6152 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1003 16:25:07.870278 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1003 16:25:07.870289 6152 ovnkube.go:599] Stopped ovnkube\\\\nI1003 16:25:07.870288 6152 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1003 16:25:07.870294 6152 handler.go:208] Removed *v1.Node event handler 7\\\\nI1003 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\" Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.572146 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81781145-978c-49ef-bb8e-4bf4f5eab778\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011b2e41b97e95ff1a11ff71d90fc00813ab18eac48b6cc7a987039494649a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6b1c34f972682f8afe389a96e2b35b848e4ceb02cb9379fa91c764f2cf11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52cbfb9a1c58ce525029b4226599a8e258453a8c5837643a6fd1efe2441641a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.577297 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.577398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.577427 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.577459 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.577482 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:33Z","lastTransitionTime":"2025-10-03T16:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.591967 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.609945 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:33Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.680561 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.680636 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.680650 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.680674 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.680687 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:33Z","lastTransitionTime":"2025-10-03T16:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.783122 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.783178 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.783195 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.783222 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.783242 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:33Z","lastTransitionTime":"2025-10-03T16:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.886693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.886749 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.886760 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.886783 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.886795 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:33Z","lastTransitionTime":"2025-10-03T16:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.891229 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.891338 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.891397 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:33 crc kubenswrapper[4744]: E1003 16:25:33.891436 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.891255 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:33 crc kubenswrapper[4744]: E1003 16:25:33.891622 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:33 crc kubenswrapper[4744]: E1003 16:25:33.891728 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:33 crc kubenswrapper[4744]: E1003 16:25:33.891824 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.990159 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.990240 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.990262 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.990295 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:33 crc kubenswrapper[4744]: I1003 16:25:33.990318 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:33Z","lastTransitionTime":"2025-10-03T16:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.093302 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.093399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.093419 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.093452 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.093472 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:34Z","lastTransitionTime":"2025-10-03T16:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.197344 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.197409 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.197426 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.197447 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.197462 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:34Z","lastTransitionTime":"2025-10-03T16:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.277164 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovnkube-controller/2.log" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.282543 4744 scope.go:117] "RemoveContainer" containerID="ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc" Oct 03 16:25:34 crc kubenswrapper[4744]: E1003 16:25:34.282864 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.298911 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.300769 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.300889 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.300968 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.301080 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.301166 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:34Z","lastTransitionTime":"2025-10-03T16:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.314510 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.352167 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\" Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.379424 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81781145-978c-49ef-bb8e-4bf4f5eab778\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011b2e41b97e95ff1a11ff71d90fc00813ab18eac48b6cc7a987039494649a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6b1c34f972682f8afe389a96e2b35b848e4ceb02cb9379fa91c764f2cf11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52cbfb9a1c58ce525029b4226599a8e258453a8c5837643a6fd1efe2441641a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.399662 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.403434 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.403613 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.403728 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.403797 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.403865 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:34Z","lastTransitionTime":"2025-10-03T16:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.417945 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.431661 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.445373 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.462571 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.479910 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.493227 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.502434 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.505927 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.506052 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.506148 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.506227 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.506301 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:34Z","lastTransitionTime":"2025-10-03T16:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.513421 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.530223 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.542528 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.573802 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.593176 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.609666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.609716 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.609731 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.609752 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.609763 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:34Z","lastTransitionTime":"2025-10-03T16:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.611752 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.712750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.712839 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.712864 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.712898 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.712918 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:34Z","lastTransitionTime":"2025-10-03T16:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.815590 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.815658 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.815670 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.815697 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.815713 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:34Z","lastTransitionTime":"2025-10-03T16:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.906672 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.918634 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.918671 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.918680 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.918697 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.918709 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:34Z","lastTransitionTime":"2025-10-03T16:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.925657 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.943392 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.961612 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:34 crc kubenswrapper[4744]: I1003 16:25:34.976933 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:34Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.004306 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:35Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.020615 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:35Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.021828 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.021865 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.021907 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.021932 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.021948 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:35Z","lastTransitionTime":"2025-10-03T16:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.035931 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:35Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.053107 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:35Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.065538 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:35Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.091317 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\" Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:35Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.105992 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81781145-978c-49ef-bb8e-4bf4f5eab778\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011b2e41b97e95ff1a11ff71d90fc00813ab18eac48b6cc7a987039494649a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6b1c34f972682f8afe389a96e2b35b848e4ceb02cb9379fa91c764f2cf11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52cbfb9a1c58ce525029b4226599a8e258453a8c5837643a6fd1efe2441641a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:35Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.124284 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.124318 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.124327 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.124342 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.124352 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:35Z","lastTransitionTime":"2025-10-03T16:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.133366 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:35Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.153702 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:35Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.168968 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:35Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.181729 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:35Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.199433 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:35Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.216208 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:35Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.228304 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.228380 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.228402 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.228435 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.228464 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:35Z","lastTransitionTime":"2025-10-03T16:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.331662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.331704 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.331718 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.331739 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.331753 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:35Z","lastTransitionTime":"2025-10-03T16:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.434596 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.434655 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.434668 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.434692 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.434707 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:35Z","lastTransitionTime":"2025-10-03T16:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.538400 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.538459 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.538470 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.538503 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.538517 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:35Z","lastTransitionTime":"2025-10-03T16:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.642951 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.643079 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.643098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.643124 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.643144 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:35Z","lastTransitionTime":"2025-10-03T16:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.747057 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.747120 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.747132 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.747153 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.747169 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:35Z","lastTransitionTime":"2025-10-03T16:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.851086 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.851163 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.851183 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.851218 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.851238 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:35Z","lastTransitionTime":"2025-10-03T16:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.890980 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.891084 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:35 crc kubenswrapper[4744]: E1003 16:25:35.891190 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.891107 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.891260 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:35 crc kubenswrapper[4744]: E1003 16:25:35.891298 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:35 crc kubenswrapper[4744]: E1003 16:25:35.891606 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:35 crc kubenswrapper[4744]: E1003 16:25:35.891728 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.955038 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.955101 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.955115 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.955138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:35 crc kubenswrapper[4744]: I1003 16:25:35.955154 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:35Z","lastTransitionTime":"2025-10-03T16:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.059274 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.059328 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.059367 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.059403 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.059431 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:36Z","lastTransitionTime":"2025-10-03T16:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.162662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.162723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.162746 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.162772 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.162807 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:36Z","lastTransitionTime":"2025-10-03T16:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.266057 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.266133 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.266155 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.266190 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.266217 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:36Z","lastTransitionTime":"2025-10-03T16:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.368604 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.368641 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.368650 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.368663 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.368673 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:36Z","lastTransitionTime":"2025-10-03T16:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.471915 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.471975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.471985 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.472004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.472016 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:36Z","lastTransitionTime":"2025-10-03T16:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.575376 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.575440 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.575454 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.575517 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.575534 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:36Z","lastTransitionTime":"2025-10-03T16:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.678536 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.678592 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.678603 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.678621 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.678632 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:36Z","lastTransitionTime":"2025-10-03T16:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.782178 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.782228 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.782238 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.782254 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.782264 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:36Z","lastTransitionTime":"2025-10-03T16:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.885633 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.885683 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.885695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.885714 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.885724 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:36Z","lastTransitionTime":"2025-10-03T16:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.988863 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.988907 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.988916 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.988931 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:36 crc kubenswrapper[4744]: I1003 16:25:36.988941 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:36Z","lastTransitionTime":"2025-10-03T16:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.091524 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.091630 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.091649 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.091664 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.091676 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:37Z","lastTransitionTime":"2025-10-03T16:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.194189 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.194244 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.194257 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.194275 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.194286 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:37Z","lastTransitionTime":"2025-10-03T16:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.296706 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.296750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.296761 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.296776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.296786 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:37Z","lastTransitionTime":"2025-10-03T16:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.399967 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.400051 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.400061 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.400078 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.400089 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:37Z","lastTransitionTime":"2025-10-03T16:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.502819 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.502871 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.502888 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.502915 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.502973 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:37Z","lastTransitionTime":"2025-10-03T16:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.606171 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.606220 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.606233 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.606253 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.606270 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:37Z","lastTransitionTime":"2025-10-03T16:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.714459 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.714548 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.714565 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.714594 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.714612 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:37Z","lastTransitionTime":"2025-10-03T16:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.817303 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.817708 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.817808 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.817906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.818030 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:37Z","lastTransitionTime":"2025-10-03T16:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.890928 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.890972 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.891029 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.891071 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:37 crc kubenswrapper[4744]: E1003 16:25:37.891330 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:37 crc kubenswrapper[4744]: E1003 16:25:37.891936 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:37 crc kubenswrapper[4744]: E1003 16:25:37.892059 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:37 crc kubenswrapper[4744]: E1003 16:25:37.892125 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.920377 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.920820 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.920902 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.920989 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:37 crc kubenswrapper[4744]: I1003 16:25:37.921099 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:37Z","lastTransitionTime":"2025-10-03T16:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.024569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.024614 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.024625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.024642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.024653 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:38Z","lastTransitionTime":"2025-10-03T16:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.126618 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.127222 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.127304 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.127405 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.127473 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:38Z","lastTransitionTime":"2025-10-03T16:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.231036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.231126 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.231137 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.231156 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.231166 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:38Z","lastTransitionTime":"2025-10-03T16:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.333757 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.333821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.333833 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.333854 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.333866 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:38Z","lastTransitionTime":"2025-10-03T16:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.436980 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.437033 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.437048 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.437069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.437082 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:38Z","lastTransitionTime":"2025-10-03T16:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.541016 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.541056 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.541065 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.541077 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.541085 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:38Z","lastTransitionTime":"2025-10-03T16:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.644060 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.644105 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.644116 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.644136 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.644147 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:38Z","lastTransitionTime":"2025-10-03T16:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.747059 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.747114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.747131 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.747158 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.747178 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:38Z","lastTransitionTime":"2025-10-03T16:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.850417 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.850859 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.850873 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.850896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.850912 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:38Z","lastTransitionTime":"2025-10-03T16:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.953977 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.954021 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.954030 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.954048 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:38 crc kubenswrapper[4744]: I1003 16:25:38.954089 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:38Z","lastTransitionTime":"2025-10-03T16:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.056922 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.056962 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.056972 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.056988 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.056999 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:39Z","lastTransitionTime":"2025-10-03T16:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.159458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.159525 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.159536 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.159553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.159568 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:39Z","lastTransitionTime":"2025-10-03T16:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.263155 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.263206 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.263218 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.263237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.263250 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:39Z","lastTransitionTime":"2025-10-03T16:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.366069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.366121 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.366132 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.366366 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.366394 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:39Z","lastTransitionTime":"2025-10-03T16:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.469536 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.469598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.469607 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.469623 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.469632 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:39Z","lastTransitionTime":"2025-10-03T16:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.572150 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.572179 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.572188 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.572204 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.572213 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:39Z","lastTransitionTime":"2025-10-03T16:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.675947 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.676022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.676033 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.676053 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.676068 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:39Z","lastTransitionTime":"2025-10-03T16:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.779343 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.779409 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.779418 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.779434 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.779445 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:39Z","lastTransitionTime":"2025-10-03T16:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.882953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.883013 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.883028 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.883058 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.883075 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:39Z","lastTransitionTime":"2025-10-03T16:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.891223 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.891302 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.891233 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.891233 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:39 crc kubenswrapper[4744]: E1003 16:25:39.891443 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:39 crc kubenswrapper[4744]: E1003 16:25:39.891576 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:39 crc kubenswrapper[4744]: E1003 16:25:39.891705 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:39 crc kubenswrapper[4744]: E1003 16:25:39.891897 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.986037 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.986081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.986094 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.986115 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:39 crc kubenswrapper[4744]: I1003 16:25:39.986130 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:39Z","lastTransitionTime":"2025-10-03T16:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.089081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.089140 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.089152 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.089168 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.089182 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:40Z","lastTransitionTime":"2025-10-03T16:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.191661 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.191723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.191736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.191756 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.191765 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:40Z","lastTransitionTime":"2025-10-03T16:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.294739 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.294794 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.294806 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.294823 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.294836 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:40Z","lastTransitionTime":"2025-10-03T16:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.397430 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.397471 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.397482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.397523 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.397536 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:40Z","lastTransitionTime":"2025-10-03T16:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.500373 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.500429 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.500444 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.500470 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.500483 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:40Z","lastTransitionTime":"2025-10-03T16:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.603015 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.603063 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.603073 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.603091 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.603101 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:40Z","lastTransitionTime":"2025-10-03T16:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.706857 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.706905 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.706917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.706939 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.706952 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:40Z","lastTransitionTime":"2025-10-03T16:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.810686 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.810751 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.810767 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.810791 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.810904 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:40Z","lastTransitionTime":"2025-10-03T16:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.913822 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.913861 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.913871 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.913885 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:40 crc kubenswrapper[4744]: I1003 16:25:40.913894 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:40Z","lastTransitionTime":"2025-10-03T16:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.016172 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.016230 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.016243 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.016261 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.016274 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:41Z","lastTransitionTime":"2025-10-03T16:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.143945 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.143988 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.143998 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.144013 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.144024 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:41Z","lastTransitionTime":"2025-10-03T16:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.246545 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.246593 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.246604 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.246619 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.246630 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:41Z","lastTransitionTime":"2025-10-03T16:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.348395 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.348439 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.348448 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.348463 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.348474 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:41Z","lastTransitionTime":"2025-10-03T16:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.450928 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.451000 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.451010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.451028 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.451040 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:41Z","lastTransitionTime":"2025-10-03T16:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.553065 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.553105 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.553116 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.553129 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.553140 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:41Z","lastTransitionTime":"2025-10-03T16:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.657036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.657107 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.657121 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.657142 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.657158 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:41Z","lastTransitionTime":"2025-10-03T16:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.761085 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.761151 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.761170 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.761198 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.761220 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:41Z","lastTransitionTime":"2025-10-03T16:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.862456 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs\") pod \"network-metrics-daemon-4f7dk\" (UID: \"1008bf15-63d7-45f4-b42b-a4680b7dc232\") " pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:41 crc kubenswrapper[4744]: E1003 16:25:41.862776 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:25:41 crc kubenswrapper[4744]: E1003 16:25:41.862869 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs podName:1008bf15-63d7-45f4-b42b-a4680b7dc232 nodeName:}" failed. No retries permitted until 2025-10-03 16:26:13.862847618 +0000 UTC m=+100.142723514 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs") pod "network-metrics-daemon-4f7dk" (UID: "1008bf15-63d7-45f4-b42b-a4680b7dc232") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.864237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.864270 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.864286 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.864313 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.864332 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:41Z","lastTransitionTime":"2025-10-03T16:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.891312 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.891357 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.891358 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.891344 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:41 crc kubenswrapper[4744]: E1003 16:25:41.891635 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:41 crc kubenswrapper[4744]: E1003 16:25:41.891766 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:41 crc kubenswrapper[4744]: E1003 16:25:41.891872 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:41 crc kubenswrapper[4744]: E1003 16:25:41.891935 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.967110 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.967549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.967774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.967966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:41 crc kubenswrapper[4744]: I1003 16:25:41.968152 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:41Z","lastTransitionTime":"2025-10-03T16:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.071166 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.071568 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.071681 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.071775 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.071875 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:42Z","lastTransitionTime":"2025-10-03T16:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.175081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.175339 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.175439 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.175527 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.175614 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:42Z","lastTransitionTime":"2025-10-03T16:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.278774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.279430 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.279534 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.279625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.279704 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:42Z","lastTransitionTime":"2025-10-03T16:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.382760 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.382841 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.382871 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.382903 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.382930 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:42Z","lastTransitionTime":"2025-10-03T16:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.486921 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.486994 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.487004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.487022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.487032 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:42Z","lastTransitionTime":"2025-10-03T16:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.591582 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.591631 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.591644 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.591673 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.591688 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:42Z","lastTransitionTime":"2025-10-03T16:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.672864 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.672918 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.672929 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.672945 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.672954 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:42Z","lastTransitionTime":"2025-10-03T16:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:42 crc kubenswrapper[4744]: E1003 16:25:42.696248 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:42Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.701584 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.701629 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.701644 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.701669 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.701685 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:42Z","lastTransitionTime":"2025-10-03T16:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:42 crc kubenswrapper[4744]: E1003 16:25:42.716355 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:42Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.720884 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.720938 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.720953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.720975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.720988 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:42Z","lastTransitionTime":"2025-10-03T16:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:42 crc kubenswrapper[4744]: E1003 16:25:42.736667 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:42Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.740925 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.740968 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.740979 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.740999 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.741025 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:42Z","lastTransitionTime":"2025-10-03T16:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:42 crc kubenswrapper[4744]: E1003 16:25:42.754577 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:42Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.759017 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.759064 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.759078 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.759098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.759112 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:42Z","lastTransitionTime":"2025-10-03T16:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:42 crc kubenswrapper[4744]: E1003 16:25:42.772406 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:42Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:42 crc kubenswrapper[4744]: E1003 16:25:42.772588 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.774832 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.774878 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.774893 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.774909 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.775020 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:42Z","lastTransitionTime":"2025-10-03T16:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.878041 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.878079 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.878091 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.878136 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.878152 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:42Z","lastTransitionTime":"2025-10-03T16:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.981776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.981842 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.981860 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.982237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:42 crc kubenswrapper[4744]: I1003 16:25:42.982277 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:42Z","lastTransitionTime":"2025-10-03T16:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.084884 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.084922 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.084934 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.084957 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.084978 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:43Z","lastTransitionTime":"2025-10-03T16:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.187882 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.187932 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.187946 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.187967 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.187980 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:43Z","lastTransitionTime":"2025-10-03T16:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.291473 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.291543 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.291560 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.291581 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.291596 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:43Z","lastTransitionTime":"2025-10-03T16:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.394595 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.394707 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.394723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.394748 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.394765 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:43Z","lastTransitionTime":"2025-10-03T16:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.497930 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.498004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.498024 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.498056 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.498078 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:43Z","lastTransitionTime":"2025-10-03T16:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.601538 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.602010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.602112 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.602221 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.602313 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:43Z","lastTransitionTime":"2025-10-03T16:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.705809 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.705876 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.705890 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.705915 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.705930 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:43Z","lastTransitionTime":"2025-10-03T16:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.809832 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.810435 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.810692 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.810939 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.811137 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:43Z","lastTransitionTime":"2025-10-03T16:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.891566 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.891666 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.892220 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.892411 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:43 crc kubenswrapper[4744]: E1003 16:25:43.892413 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:43 crc kubenswrapper[4744]: E1003 16:25:43.892625 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:43 crc kubenswrapper[4744]: E1003 16:25:43.892822 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:43 crc kubenswrapper[4744]: E1003 16:25:43.893040 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.915464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.915574 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.915599 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.915667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:43 crc kubenswrapper[4744]: I1003 16:25:43.915692 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:43Z","lastTransitionTime":"2025-10-03T16:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.019721 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.020240 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.020437 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.020689 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.020881 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:44Z","lastTransitionTime":"2025-10-03T16:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.124876 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.124922 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.124952 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.124976 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.124991 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:44Z","lastTransitionTime":"2025-10-03T16:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.228174 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.228461 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.228571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.228662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.228732 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:44Z","lastTransitionTime":"2025-10-03T16:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.316149 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vqz6q_84192d79-366a-453f-b70b-aefa4537ec4c/kube-multus/0.log" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.316769 4744 generic.go:334] "Generic (PLEG): container finished" podID="84192d79-366a-453f-b70b-aefa4537ec4c" containerID="c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17" exitCode=1 Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.316823 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vqz6q" event={"ID":"84192d79-366a-453f-b70b-aefa4537ec4c","Type":"ContainerDied","Data":"c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17"} Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.317608 4744 scope.go:117] "RemoveContainer" containerID="c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.332762 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.332803 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.332814 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.332831 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.332841 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:44Z","lastTransitionTime":"2025-10-03T16:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.337637 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.360909 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\" Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.378548 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81781145-978c-49ef-bb8e-4bf4f5eab778\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011b2e41b97e95ff1a11ff71d90fc00813ab18eac48b6cc7a987039494649a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6b1c34f972682f8afe389a96e2b35b848e4ceb02cb9379fa91c764f2cf11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52cbfb9a1c58ce525029b4226599a8e258453a8c5837643a6fd1efe2441641a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.392564 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.411845 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.429469 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.437985 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.438035 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.438046 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.438063 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.438077 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:44Z","lastTransitionTime":"2025-10-03T16:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.446333 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:44Z\\\",\\\"message\\\":\\\"2025-10-03T16:24:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41\\\\n2025-10-03T16:24:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41 to /host/opt/cni/bin/\\\\n2025-10-03T16:24:59Z [verbose] multus-daemon started\\\\n2025-10-03T16:24:59Z [verbose] Readiness Indicator file check\\\\n2025-10-03T16:25:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.462164 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.476194 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.489889 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.500934 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.513703 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.524571 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.534029 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.540910 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.540964 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.540984 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.541015 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.541035 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:44Z","lastTransitionTime":"2025-10-03T16:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.552565 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.564991 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.581123 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.595684 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.644623 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.644662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.644671 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.644717 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.644730 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:44Z","lastTransitionTime":"2025-10-03T16:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.748245 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.748339 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.748350 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.748368 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.748380 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:44Z","lastTransitionTime":"2025-10-03T16:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.851267 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.851325 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.851365 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.851381 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.851389 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:44Z","lastTransitionTime":"2025-10-03T16:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.906569 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.927739 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:44Z\\\",\\\"message\\\":\\\"2025-10-03T16:24:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41\\\\n2025-10-03T16:24:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41 to /host/opt/cni/bin/\\\\n2025-10-03T16:24:59Z [verbose] multus-daemon started\\\\n2025-10-03T16:24:59Z [verbose] Readiness Indicator file check\\\\n2025-10-03T16:25:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.942404 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.955812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.955863 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.955875 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.955893 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.955929 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:44Z","lastTransitionTime":"2025-10-03T16:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.972568 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\" Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:44 crc kubenswrapper[4744]: I1003 16:25:44.989286 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81781145-978c-49ef-bb8e-4bf4f5eab778\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011b2e41b97e95ff1a11ff71d90fc00813ab18eac48b6cc7a987039494649a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6b1c34f972682f8afe389a96e2b35b848e4ceb02cb9379fa91c764f2cf11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52cbfb9a1c58ce525029b4226599a8e258453a8c5837643a6fd1efe2441641a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:44Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.003816 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.018321 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.031179 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.043316 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.057539 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.058193 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.058226 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.058238 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.058256 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.058267 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:45Z","lastTransitionTime":"2025-10-03T16:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.072330 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.087025 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.104384 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.121568 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.139896 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.155127 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.160743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.160781 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.160792 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.160811 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.160823 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:45Z","lastTransitionTime":"2025-10-03T16:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.180594 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.196531 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.264776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.264832 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.264849 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.264880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.264901 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:45Z","lastTransitionTime":"2025-10-03T16:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.322946 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vqz6q_84192d79-366a-453f-b70b-aefa4537ec4c/kube-multus/0.log" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.323010 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vqz6q" event={"ID":"84192d79-366a-453f-b70b-aefa4537ec4c","Type":"ContainerStarted","Data":"d3888e7ce3b4e4e7967e3c7ec2e5ef8b7bee43a7742087ca6ea6f44a3d72e45c"} Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.350729 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.367933 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.367987 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.368000 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.368018 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.368034 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:45Z","lastTransitionTime":"2025-10-03T16:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.373390 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.390399 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.403271 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.414992 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.428152 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.448342 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.463601 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.471007 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.471045 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.471057 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.471076 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.471088 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:45Z","lastTransitionTime":"2025-10-03T16:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.477871 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.496173 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.511702 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.531756 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\" Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.543024 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81781145-978c-49ef-bb8e-4bf4f5eab778\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011b2e41b97e95ff1a11ff71d90fc00813ab18eac48b6cc7a987039494649a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6b1c34f972682f8afe389a96e2b35b848e4ceb02cb9379fa91c764f2cf11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52cbfb9a1c58ce525029b4226599a8e258453a8c5837643a6fd1efe2441641a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.555345 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.570589 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.573615 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.573696 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.573742 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.573775 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.573789 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:45Z","lastTransitionTime":"2025-10-03T16:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.585563 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.598603 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3888e7ce3b4e4e7967e3c7ec2e5ef8b7bee43a7742087ca6ea6f44a3d72e45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:44Z\\\",\\\"message\\\":\\\"2025-10-03T16:24:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41\\\\n2025-10-03T16:24:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41 to /host/opt/cni/bin/\\\\n2025-10-03T16:24:59Z [verbose] multus-daemon started\\\\n2025-10-03T16:24:59Z [verbose] Readiness Indicator file check\\\\n2025-10-03T16:25:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.608457 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:45Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.676730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.676801 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.676821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.676847 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.676867 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:45Z","lastTransitionTime":"2025-10-03T16:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.780616 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.780676 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.780692 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.780716 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.780733 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:45Z","lastTransitionTime":"2025-10-03T16:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.884413 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.884482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.884508 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.884525 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.884535 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:45Z","lastTransitionTime":"2025-10-03T16:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.891826 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.891878 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:45 crc kubenswrapper[4744]: E1003 16:25:45.891955 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.891832 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.891897 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:45 crc kubenswrapper[4744]: E1003 16:25:45.892095 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:45 crc kubenswrapper[4744]: E1003 16:25:45.892220 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:45 crc kubenswrapper[4744]: E1003 16:25:45.892290 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.987714 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.987780 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.987789 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.987811 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:45 crc kubenswrapper[4744]: I1003 16:25:45.987843 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:45Z","lastTransitionTime":"2025-10-03T16:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.090397 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.090703 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.090788 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.090877 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.090970 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:46Z","lastTransitionTime":"2025-10-03T16:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.194026 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.194077 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.194089 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.194106 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.194120 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:46Z","lastTransitionTime":"2025-10-03T16:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.298259 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.298322 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.298335 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.298357 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.298400 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:46Z","lastTransitionTime":"2025-10-03T16:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.400842 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.400893 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.400904 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.400922 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.400934 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:46Z","lastTransitionTime":"2025-10-03T16:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.503843 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.503895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.503911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.503930 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.503944 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:46Z","lastTransitionTime":"2025-10-03T16:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.606431 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.606481 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.606505 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.606540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.606558 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:46Z","lastTransitionTime":"2025-10-03T16:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.709994 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.710098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.710126 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.710162 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.710193 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:46Z","lastTransitionTime":"2025-10-03T16:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.813346 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.813408 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.813434 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.813485 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.813575 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:46Z","lastTransitionTime":"2025-10-03T16:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.894627 4744 scope.go:117] "RemoveContainer" containerID="ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc" Oct 03 16:25:46 crc kubenswrapper[4744]: E1003 16:25:46.895909 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.916875 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.916932 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.916944 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.916964 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:46 crc kubenswrapper[4744]: I1003 16:25:46.916980 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:46Z","lastTransitionTime":"2025-10-03T16:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.019751 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.019795 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.019806 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.019824 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.019836 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:47Z","lastTransitionTime":"2025-10-03T16:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.122693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.122775 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.122794 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.122823 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.122847 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:47Z","lastTransitionTime":"2025-10-03T16:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.226804 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.226860 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.226871 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.226897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.226911 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:47Z","lastTransitionTime":"2025-10-03T16:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.328966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.329023 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.329042 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.329072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.329094 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:47Z","lastTransitionTime":"2025-10-03T16:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.431440 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.431525 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.431539 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.431554 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.431566 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:47Z","lastTransitionTime":"2025-10-03T16:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.533841 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.533898 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.533911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.533931 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.533947 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:47Z","lastTransitionTime":"2025-10-03T16:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.637048 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.637120 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.637139 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.637169 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.637189 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:47Z","lastTransitionTime":"2025-10-03T16:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.739991 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.740068 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.740089 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.740117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.740137 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:47Z","lastTransitionTime":"2025-10-03T16:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.843027 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.843086 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.843105 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.843137 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.843157 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:47Z","lastTransitionTime":"2025-10-03T16:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.891342 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:47 crc kubenswrapper[4744]: E1003 16:25:47.891487 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.891668 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.891725 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:47 crc kubenswrapper[4744]: E1003 16:25:47.891976 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.891709 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:47 crc kubenswrapper[4744]: E1003 16:25:47.892212 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:47 crc kubenswrapper[4744]: E1003 16:25:47.892224 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.946835 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.946914 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.946927 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.946954 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:47 crc kubenswrapper[4744]: I1003 16:25:47.946974 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:47Z","lastTransitionTime":"2025-10-03T16:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.050629 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.050695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.050710 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.050732 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.050745 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:48Z","lastTransitionTime":"2025-10-03T16:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.153589 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.153642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.153651 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.153671 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.153683 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:48Z","lastTransitionTime":"2025-10-03T16:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.257308 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.257358 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.257368 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.257386 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.257398 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:48Z","lastTransitionTime":"2025-10-03T16:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.360549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.360598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.360608 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.360624 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.360634 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:48Z","lastTransitionTime":"2025-10-03T16:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.464216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.464277 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.464296 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.464324 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.464341 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:48Z","lastTransitionTime":"2025-10-03T16:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.567571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.567669 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.567692 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.567725 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.567748 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:48Z","lastTransitionTime":"2025-10-03T16:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.671086 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.671598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.671851 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.672026 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.672167 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:48Z","lastTransitionTime":"2025-10-03T16:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.775702 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.775771 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.775788 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.775812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.775834 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:48Z","lastTransitionTime":"2025-10-03T16:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.879323 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.879398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.879411 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.879434 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.879451 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:48Z","lastTransitionTime":"2025-10-03T16:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.981662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.982087 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.982104 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.982126 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:48 crc kubenswrapper[4744]: I1003 16:25:48.982144 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:48Z","lastTransitionTime":"2025-10-03T16:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.085830 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.086209 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.086327 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.086409 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.086510 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:49Z","lastTransitionTime":"2025-10-03T16:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.189982 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.190068 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.190094 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.190132 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.190159 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:49Z","lastTransitionTime":"2025-10-03T16:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.292762 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.292836 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.292856 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.292890 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.292914 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:49Z","lastTransitionTime":"2025-10-03T16:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.397123 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.397629 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.397806 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.397948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.398074 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:49Z","lastTransitionTime":"2025-10-03T16:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.501227 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.501288 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.501300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.501325 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.501339 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:49Z","lastTransitionTime":"2025-10-03T16:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.603209 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.603256 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.603265 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.603277 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.603286 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:49Z","lastTransitionTime":"2025-10-03T16:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.707126 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.707187 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.707209 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.707239 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.707259 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:49Z","lastTransitionTime":"2025-10-03T16:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.810679 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.810759 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.810786 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.810823 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.810851 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:49Z","lastTransitionTime":"2025-10-03T16:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.891104 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.891108 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.891187 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.891310 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:49 crc kubenswrapper[4744]: E1003 16:25:49.891599 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:49 crc kubenswrapper[4744]: E1003 16:25:49.891827 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:49 crc kubenswrapper[4744]: E1003 16:25:49.891995 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:49 crc kubenswrapper[4744]: E1003 16:25:49.892075 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.913885 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.913942 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.913972 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.914001 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:49 crc kubenswrapper[4744]: I1003 16:25:49.914019 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:49Z","lastTransitionTime":"2025-10-03T16:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.017130 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.017218 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.017244 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.017285 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.017312 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:50Z","lastTransitionTime":"2025-10-03T16:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.120349 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.120410 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.120422 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.120439 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.120451 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:50Z","lastTransitionTime":"2025-10-03T16:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.223629 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.224031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.224199 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.224337 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.224463 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:50Z","lastTransitionTime":"2025-10-03T16:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.328825 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.329275 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.329564 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.329798 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.330054 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:50Z","lastTransitionTime":"2025-10-03T16:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.433563 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.433622 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.433638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.433662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.433679 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:50Z","lastTransitionTime":"2025-10-03T16:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.537717 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.537810 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.537835 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.537872 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.537899 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:50Z","lastTransitionTime":"2025-10-03T16:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.641862 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.641948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.641974 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.642010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.642038 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:50Z","lastTransitionTime":"2025-10-03T16:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.745050 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.745113 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.745125 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.745144 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.745158 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:50Z","lastTransitionTime":"2025-10-03T16:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.848010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.848074 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.848091 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.848114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.848129 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:50Z","lastTransitionTime":"2025-10-03T16:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.951783 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.951848 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.951861 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.951880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:50 crc kubenswrapper[4744]: I1003 16:25:50.952246 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:50Z","lastTransitionTime":"2025-10-03T16:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.055647 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.055739 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.055768 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.055803 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.055831 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:51Z","lastTransitionTime":"2025-10-03T16:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.158085 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.158135 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.158147 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.158166 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.158177 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:51Z","lastTransitionTime":"2025-10-03T16:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.261273 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.261353 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.261380 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.261398 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.261409 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:51Z","lastTransitionTime":"2025-10-03T16:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.363643 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.363953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.364129 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.364247 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.364361 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:51Z","lastTransitionTime":"2025-10-03T16:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.467352 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.467430 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.467444 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.467468 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.467482 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:51Z","lastTransitionTime":"2025-10-03T16:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.570881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.570954 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.570967 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.570996 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.571014 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:51Z","lastTransitionTime":"2025-10-03T16:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.674151 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.674201 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.674213 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.674227 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.674236 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:51Z","lastTransitionTime":"2025-10-03T16:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.776909 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.776968 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.776984 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.777009 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.777026 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:51Z","lastTransitionTime":"2025-10-03T16:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.880040 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.880372 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.880473 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.880600 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.880682 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:51Z","lastTransitionTime":"2025-10-03T16:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.891375 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.891461 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.891468 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:51 crc kubenswrapper[4744]: E1003 16:25:51.892038 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:51 crc kubenswrapper[4744]: E1003 16:25:51.891849 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:51 crc kubenswrapper[4744]: E1003 16:25:51.892103 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.891614 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:51 crc kubenswrapper[4744]: E1003 16:25:51.892297 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.982923 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.983170 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.983283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.983347 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:51 crc kubenswrapper[4744]: I1003 16:25:51.983436 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:51Z","lastTransitionTime":"2025-10-03T16:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.087055 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.087420 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.087511 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.087627 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.087747 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:52Z","lastTransitionTime":"2025-10-03T16:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.190975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.191040 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.191069 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.191112 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.191136 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:52Z","lastTransitionTime":"2025-10-03T16:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.294121 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.294417 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.294533 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.294642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.294719 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:52Z","lastTransitionTime":"2025-10-03T16:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.397559 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.397994 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.398087 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.398182 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.398262 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:52Z","lastTransitionTime":"2025-10-03T16:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.501999 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.502049 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.502060 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.502077 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.502089 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:52Z","lastTransitionTime":"2025-10-03T16:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.605250 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.605299 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.605310 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.605332 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.605349 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:52Z","lastTransitionTime":"2025-10-03T16:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.709277 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.709343 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.709366 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.709400 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.709431 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:52Z","lastTransitionTime":"2025-10-03T16:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.777484 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.777656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.777679 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.777713 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.777740 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:52Z","lastTransitionTime":"2025-10-03T16:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:52 crc kubenswrapper[4744]: E1003 16:25:52.800851 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:52Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.806113 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.806153 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.806167 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.806188 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.806205 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:52Z","lastTransitionTime":"2025-10-03T16:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:52 crc kubenswrapper[4744]: E1003 16:25:52.825560 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:52Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.830569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.830605 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.830619 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.830639 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.830651 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:52Z","lastTransitionTime":"2025-10-03T16:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:52 crc kubenswrapper[4744]: E1003 16:25:52.849390 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:52Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.855539 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.855620 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.855644 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.855676 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.855699 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:52Z","lastTransitionTime":"2025-10-03T16:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:52 crc kubenswrapper[4744]: E1003 16:25:52.873071 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:52Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.879827 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.879891 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.879906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.879928 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.879965 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:52Z","lastTransitionTime":"2025-10-03T16:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:52 crc kubenswrapper[4744]: E1003 16:25:52.897238 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:52Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:52 crc kubenswrapper[4744]: E1003 16:25:52.897559 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.899831 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.899951 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.900057 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.900154 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:52 crc kubenswrapper[4744]: I1003 16:25:52.900262 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:52Z","lastTransitionTime":"2025-10-03T16:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.004590 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.004641 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.004663 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.004709 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.004730 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:53Z","lastTransitionTime":"2025-10-03T16:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.109900 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.110238 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.110359 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.110464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.110583 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:53Z","lastTransitionTime":"2025-10-03T16:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.213913 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.213987 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.214005 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.214032 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.214050 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:53Z","lastTransitionTime":"2025-10-03T16:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.317584 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.317647 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.317666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.317691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.317711 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:53Z","lastTransitionTime":"2025-10-03T16:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.420980 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.421040 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.421060 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.421090 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.421111 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:53Z","lastTransitionTime":"2025-10-03T16:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.524431 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.524520 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.524535 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.524556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.524570 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:53Z","lastTransitionTime":"2025-10-03T16:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.627541 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.627616 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.627632 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.627653 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.627666 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:53Z","lastTransitionTime":"2025-10-03T16:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.731292 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.731359 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.731378 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.731408 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.731428 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:53Z","lastTransitionTime":"2025-10-03T16:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.834396 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.834464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.834484 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.834541 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.834564 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:53Z","lastTransitionTime":"2025-10-03T16:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.891152 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:53 crc kubenswrapper[4744]: E1003 16:25:53.891326 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.891597 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:53 crc kubenswrapper[4744]: E1003 16:25:53.891661 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.891767 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.891810 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:53 crc kubenswrapper[4744]: E1003 16:25:53.891918 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:53 crc kubenswrapper[4744]: E1003 16:25:53.892316 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.937794 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.937842 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.937853 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.937870 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:53 crc kubenswrapper[4744]: I1003 16:25:53.937879 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:53Z","lastTransitionTime":"2025-10-03T16:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.041062 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.041116 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.041125 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.041141 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.041154 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:54Z","lastTransitionTime":"2025-10-03T16:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.144053 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.144090 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.144098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.144111 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.144119 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:54Z","lastTransitionTime":"2025-10-03T16:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.246998 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.247044 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.247057 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.247074 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.247084 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:54Z","lastTransitionTime":"2025-10-03T16:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.350174 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.350219 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.350231 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.350247 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.350260 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:54Z","lastTransitionTime":"2025-10-03T16:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.452942 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.453009 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.453021 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.453050 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.453060 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:54Z","lastTransitionTime":"2025-10-03T16:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.555428 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.555464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.555473 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.555488 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.555518 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:54Z","lastTransitionTime":"2025-10-03T16:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.658688 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.658726 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.658737 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.658750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.658759 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:54Z","lastTransitionTime":"2025-10-03T16:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.761475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.761578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.761596 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.761627 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.761651 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:54Z","lastTransitionTime":"2025-10-03T16:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.863958 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.864009 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.864018 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.864035 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.864048 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:54Z","lastTransitionTime":"2025-10-03T16:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.909761 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:54Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.931286 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:54Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.942742 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:54Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.954379 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:54Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.967247 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.967315 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.967336 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.967373 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.967406 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:54Z","lastTransitionTime":"2025-10-03T16:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.969064 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:54Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:54 crc kubenswrapper[4744]: I1003 16:25:54.983558 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:54Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.002146 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\" Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:54Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.019745 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81781145-978c-49ef-bb8e-4bf4f5eab778\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011b2e41b97e95ff1a11ff71d90fc00813ab18eac48b6cc7a987039494649a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6b1c34f972682f8afe389a96e2b35b848e4ceb02cb9379fa91c764f2cf11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52cbfb9a1c58ce525029b4226599a8e258453a8c5837643a6fd1efe2441641a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:55Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.035952 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:55Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.049824 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:55Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.061542 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:55Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.069407 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.069430 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.069438 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.069454 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.069463 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:55Z","lastTransitionTime":"2025-10-03T16:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.081220 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3888e7ce3b4e4e7967e3c7ec2e5ef8b7bee43a7742087ca6ea6f44a3d72e45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:44Z\\\",\\\"message\\\":\\\"2025-10-03T16:24:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41\\\\n2025-10-03T16:24:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41 to /host/opt/cni/bin/\\\\n2025-10-03T16:24:59Z [verbose] multus-daemon started\\\\n2025-10-03T16:24:59Z [verbose] Readiness Indicator file check\\\\n2025-10-03T16:25:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:55Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.093908 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:55Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.120781 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:55Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.138901 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:55Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.153247 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:55Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.165806 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:55Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.171661 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.171706 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.171717 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.171733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.171744 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:55Z","lastTransitionTime":"2025-10-03T16:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.178452 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:55Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.274690 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.274842 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.274875 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.274951 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.274986 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:55Z","lastTransitionTime":"2025-10-03T16:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.377784 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.377859 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.377903 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.377948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.377976 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:55Z","lastTransitionTime":"2025-10-03T16:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.481408 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.481444 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.481454 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.481474 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.481484 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:55Z","lastTransitionTime":"2025-10-03T16:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.584413 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.584804 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.584902 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.584970 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.585036 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:55Z","lastTransitionTime":"2025-10-03T16:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.689135 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.689899 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.689966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.690077 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.690160 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:55Z","lastTransitionTime":"2025-10-03T16:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.793107 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.793162 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.793171 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.793187 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.793196 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:55Z","lastTransitionTime":"2025-10-03T16:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.891776 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:55 crc kubenswrapper[4744]: E1003 16:25:55.891905 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.892081 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:55 crc kubenswrapper[4744]: E1003 16:25:55.892135 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.892226 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:55 crc kubenswrapper[4744]: E1003 16:25:55.892270 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.892361 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:55 crc kubenswrapper[4744]: E1003 16:25:55.892405 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.895050 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.895093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.895104 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.895118 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.895132 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:55Z","lastTransitionTime":"2025-10-03T16:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.997387 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.997460 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.997481 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.997551 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:55 crc kubenswrapper[4744]: I1003 16:25:55.997575 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:55Z","lastTransitionTime":"2025-10-03T16:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.100909 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.100961 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.100973 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.100998 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.101017 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:56Z","lastTransitionTime":"2025-10-03T16:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.203386 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.203480 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.203529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.203663 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.203711 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:56Z","lastTransitionTime":"2025-10-03T16:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.307428 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.307809 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.307904 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.307979 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.308052 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:56Z","lastTransitionTime":"2025-10-03T16:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.410748 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.410789 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.410797 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.410812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.410820 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:56Z","lastTransitionTime":"2025-10-03T16:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.513681 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.513723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.513732 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.513748 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.513759 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:56Z","lastTransitionTime":"2025-10-03T16:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.617039 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.617362 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.617489 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.617620 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.617704 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:56Z","lastTransitionTime":"2025-10-03T16:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.720395 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.720429 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.720440 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.720455 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.720464 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:56Z","lastTransitionTime":"2025-10-03T16:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.823588 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.823918 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.823982 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.824075 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.824153 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:56Z","lastTransitionTime":"2025-10-03T16:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.927100 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.927169 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.927189 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.927215 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:56 crc kubenswrapper[4744]: I1003 16:25:56.927234 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:56Z","lastTransitionTime":"2025-10-03T16:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.030028 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.030113 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.030133 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.030162 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.030182 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:57Z","lastTransitionTime":"2025-10-03T16:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.132941 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.133008 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.133030 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.133061 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.133085 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:57Z","lastTransitionTime":"2025-10-03T16:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.235800 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.235901 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.235918 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.235940 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.235954 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:57Z","lastTransitionTime":"2025-10-03T16:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.339865 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.339982 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.340003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.340035 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.340062 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:57Z","lastTransitionTime":"2025-10-03T16:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.443447 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.443578 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.443597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.443630 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.443647 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:57Z","lastTransitionTime":"2025-10-03T16:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.546351 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.546424 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.546444 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.546475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.546538 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:57Z","lastTransitionTime":"2025-10-03T16:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.650162 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.650214 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.650229 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.650256 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.650273 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:57Z","lastTransitionTime":"2025-10-03T16:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.753899 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.753948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.753961 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.753979 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.753988 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:57Z","lastTransitionTime":"2025-10-03T16:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.857482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.857606 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.857633 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.857668 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.857697 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:57Z","lastTransitionTime":"2025-10-03T16:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.891627 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:57 crc kubenswrapper[4744]: E1003 16:25:57.891828 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.891907 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.891918 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.891903 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:57 crc kubenswrapper[4744]: E1003 16:25:57.892005 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:57 crc kubenswrapper[4744]: E1003 16:25:57.892605 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:57 crc kubenswrapper[4744]: E1003 16:25:57.892880 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.893234 4744 scope.go:117] "RemoveContainer" containerID="ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.961212 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.961269 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.961289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.961313 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:57 crc kubenswrapper[4744]: I1003 16:25:57.961331 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:57Z","lastTransitionTime":"2025-10-03T16:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.064337 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.064381 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.064392 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.064411 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.064426 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:58Z","lastTransitionTime":"2025-10-03T16:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.167622 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.167677 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.167689 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.167710 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.167725 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:58Z","lastTransitionTime":"2025-10-03T16:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.270121 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.270166 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.270176 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.270192 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.270206 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:58Z","lastTransitionTime":"2025-10-03T16:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.373528 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.373602 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.373621 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.373651 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.373671 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:58Z","lastTransitionTime":"2025-10-03T16:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.377749 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovnkube-controller/2.log" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.382978 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerStarted","Data":"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb"} Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.383975 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.430811 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.456146 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.477113 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.477292 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.477357 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.477377 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.477406 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.477427 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:58Z","lastTransitionTime":"2025-10-03T16:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.496897 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.510971 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.533552 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81781145-978c-49ef-bb8e-4bf4f5eab778\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011b2e41b97e95ff1a11ff71d90fc00813ab18eac48b6cc7a987039494649a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6b1c34f972682f8afe389a96e2b35b848e4ceb02cb9379fa91c764f2cf11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52cbfb9a1c58ce525029b4226599a8e258453a8c5837643a6fd1efe2441641a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.547848 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.562475 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.579917 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.580000 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.580099 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.580128 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.580142 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:58Z","lastTransitionTime":"2025-10-03T16:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.581886 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.618174 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3888e7ce3b4e4e7967e3c7ec2e5ef8b7bee43a7742087ca6ea6f44a3d72e45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:44Z\\\",\\\"message\\\":\\\"2025-10-03T16:24:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41\\\\n2025-10-03T16:24:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41 to /host/opt/cni/bin/\\\\n2025-10-03T16:24:59Z [verbose] multus-daemon started\\\\n2025-10-03T16:24:59Z [verbose] Readiness Indicator file check\\\\n2025-10-03T16:25:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.642607 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.662592 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.662775 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:58 crc kubenswrapper[4744]: E1003 16:25:58.662864 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:02.662827668 +0000 UTC m=+148.942703564 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:25:58 crc kubenswrapper[4744]: E1003 16:25:58.662890 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:25:58 crc kubenswrapper[4744]: E1003 16:25:58.662959 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:27:02.66294449 +0000 UTC m=+148.942820386 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.663044 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:58 crc kubenswrapper[4744]: E1003 16:25:58.663245 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:25:58 crc kubenswrapper[4744]: E1003 16:25:58.663426 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 16:27:02.663384761 +0000 UTC m=+148.943260657 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.665417 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\" Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.682417 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.683444 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.683590 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.683625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.683676 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.683689 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:58Z","lastTransitionTime":"2025-10-03T16:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.701400 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.715227 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.727199 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.740262 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.751962 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.764569 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.764624 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:58 crc kubenswrapper[4744]: E1003 16:25:58.764796 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:25:58 crc kubenswrapper[4744]: E1003 16:25:58.764811 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 16:25:58 crc kubenswrapper[4744]: E1003 16:25:58.764848 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:25:58 crc kubenswrapper[4744]: E1003 16:25:58.764863 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:58 crc kubenswrapper[4744]: E1003 16:25:58.764917 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 16:27:02.764898961 +0000 UTC m=+149.044774857 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:58 crc kubenswrapper[4744]: E1003 16:25:58.764816 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 16:25:58 crc kubenswrapper[4744]: E1003 16:25:58.764959 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:58 crc kubenswrapper[4744]: E1003 16:25:58.765047 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 16:27:02.765028254 +0000 UTC m=+149.044904150 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.786168 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.786207 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.786217 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.786233 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.786245 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:58Z","lastTransitionTime":"2025-10-03T16:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.889913 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.889965 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.889978 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.889993 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.890002 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:58Z","lastTransitionTime":"2025-10-03T16:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.992662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.992715 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.992726 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.992747 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:58 crc kubenswrapper[4744]: I1003 16:25:58.992760 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:58Z","lastTransitionTime":"2025-10-03T16:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.096284 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.096330 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.096348 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.096367 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.096378 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:59Z","lastTransitionTime":"2025-10-03T16:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.200058 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.200128 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.200149 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.200180 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.200207 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:59Z","lastTransitionTime":"2025-10-03T16:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.303691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.303763 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.303784 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.303816 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.303839 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:59Z","lastTransitionTime":"2025-10-03T16:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.389611 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovnkube-controller/3.log" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.390874 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovnkube-controller/2.log" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.395838 4744 generic.go:334] "Generic (PLEG): container finished" podID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerID="e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb" exitCode=1 Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.395927 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerDied","Data":"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb"} Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.396002 4744 scope.go:117] "RemoveContainer" containerID="ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.397409 4744 scope.go:117] "RemoveContainer" containerID="e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb" Oct 03 16:25:59 crc kubenswrapper[4744]: E1003 16:25:59.397806 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.407950 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.408482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.408512 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.408543 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.408559 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:59Z","lastTransitionTime":"2025-10-03T16:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.419807 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3888e7ce3b4e4e7967e3c7ec2e5ef8b7bee43a7742087ca6ea6f44a3d72e45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:44Z\\\",\\\"message\\\":\\\"2025-10-03T16:24:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41\\\\n2025-10-03T16:24:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41 to /host/opt/cni/bin/\\\\n2025-10-03T16:24:59Z [verbose] multus-daemon started\\\\n2025-10-03T16:24:59Z [verbose] Readiness Indicator file check\\\\n2025-10-03T16:25:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.439995 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.468639 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba13b0775832374809b791f945dca440be4fb2aa1a06aa106320fc8871f2eddc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:32Z\\\",\\\"message\\\":\\\" Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"89fe421e-04e8-4967-ac75-77a0e6f784ef\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/marketplace-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]strin\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:58Z\\\",\\\"message\\\":\\\"object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb\\\\nI1003 16:25:58.946365 6793 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1003 16:25:58.946371 6793 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-bdrsw\\\\nI1003 16:25:58.946377 6793 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF1003 16:25:58.946375 6793 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z]\\\\nI1003 16:25:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.487937 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81781145-978c-49ef-bb8e-4bf4f5eab778\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011b2e41b97e95ff1a11ff71d90fc00813ab18eac48b6cc7a987039494649a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6b1c34f972682f8afe389a96e2b35b848e4ceb02cb9379fa91c764f2cf11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52cbfb9a1c58ce525029b4226599a8e258453a8c5837643a6fd1efe2441641a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.507472 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.512081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.512313 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.512410 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.512556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.512664 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:59Z","lastTransitionTime":"2025-10-03T16:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.525089 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.540567 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.553170 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.567914 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.586093 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.602228 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.616972 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.617318 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.617410 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.617480 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.617582 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:59Z","lastTransitionTime":"2025-10-03T16:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.617120 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.633778 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.650236 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.663606 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.686917 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.700842 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.712779 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:59Z is after 2025-08-24T17:21:41Z" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.720659 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.720719 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.720736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.720766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.720783 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:59Z","lastTransitionTime":"2025-10-03T16:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.825143 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.825208 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.825245 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.825281 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.825308 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:59Z","lastTransitionTime":"2025-10-03T16:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.891342 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.891392 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:25:59 crc kubenswrapper[4744]: E1003 16:25:59.891722 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.891549 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:25:59 crc kubenswrapper[4744]: E1003 16:25:59.892056 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.891481 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:25:59 crc kubenswrapper[4744]: E1003 16:25:59.892230 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:25:59 crc kubenswrapper[4744]: E1003 16:25:59.891867 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.928573 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.928623 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.928635 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.928652 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:25:59 crc kubenswrapper[4744]: I1003 16:25:59.928666 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:25:59Z","lastTransitionTime":"2025-10-03T16:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.031690 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.031774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.031794 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.031824 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.031844 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:00Z","lastTransitionTime":"2025-10-03T16:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.134715 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.135123 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.135193 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.135263 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.135332 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:00Z","lastTransitionTime":"2025-10-03T16:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.245292 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.245348 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.245361 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.245383 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.245397 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:00Z","lastTransitionTime":"2025-10-03T16:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.349590 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.349744 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.349761 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.349783 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.349793 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:00Z","lastTransitionTime":"2025-10-03T16:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.402896 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovnkube-controller/3.log" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.407059 4744 scope.go:117] "RemoveContainer" containerID="e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb" Oct 03 16:26:00 crc kubenswrapper[4744]: E1003 16:26:00.407332 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.426974 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.442255 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.453959 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.454403 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.454714 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.454876 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.454955 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:00Z","lastTransitionTime":"2025-10-03T16:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.456211 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.475515 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.490851 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.504545 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.519600 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.536482 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.553056 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.558976 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.559021 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.559031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.559047 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.559059 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:00Z","lastTransitionTime":"2025-10-03T16:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.570882 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.595693 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.610037 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.626158 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.641295 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3888e7ce3b4e4e7967e3c7ec2e5ef8b7bee43a7742087ca6ea6f44a3d72e45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:44Z\\\",\\\"message\\\":\\\"2025-10-03T16:24:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41\\\\n2025-10-03T16:24:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41 to /host/opt/cni/bin/\\\\n2025-10-03T16:24:59Z [verbose] multus-daemon started\\\\n2025-10-03T16:24:59Z [verbose] Readiness Indicator file check\\\\n2025-10-03T16:25:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.655617 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.661340 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.661576 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.661689 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.661790 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.661883 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:00Z","lastTransitionTime":"2025-10-03T16:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.676073 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:58Z\\\",\\\"message\\\":\\\"object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb\\\\nI1003 16:25:58.946365 6793 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1003 16:25:58.946371 6793 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-bdrsw\\\\nI1003 16:25:58.946377 6793 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF1003 16:25:58.946375 6793 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z]\\\\nI1003 16:25:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.689389 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81781145-978c-49ef-bb8e-4bf4f5eab778\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011b2e41b97e95ff1a11ff71d90fc00813ab18eac48b6cc7a987039494649a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6b1c34f972682f8afe389a96e2b35b848e4ceb02cb9379fa91c764f2cf11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52cbfb9a1c58ce525029b4226599a8e258453a8c5837643a6fd1efe2441641a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.704444 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:00Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.765584 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.766935 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.766969 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.766996 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.767012 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:00Z","lastTransitionTime":"2025-10-03T16:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.869096 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.869143 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.869151 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.869165 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.869176 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:00Z","lastTransitionTime":"2025-10-03T16:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.971827 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.971884 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.971896 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.971911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:00 crc kubenswrapper[4744]: I1003 16:26:00.971923 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:00Z","lastTransitionTime":"2025-10-03T16:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.075068 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.075115 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.075127 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.075145 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.075156 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:01Z","lastTransitionTime":"2025-10-03T16:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.178399 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.178446 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.178456 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.178474 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.178484 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:01Z","lastTransitionTime":"2025-10-03T16:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.280906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.280975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.280993 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.281019 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.281037 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:01Z","lastTransitionTime":"2025-10-03T16:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.384987 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.385053 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.385073 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.385100 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.385117 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:01Z","lastTransitionTime":"2025-10-03T16:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.487250 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.487290 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.487300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.487317 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.487330 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:01Z","lastTransitionTime":"2025-10-03T16:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.590887 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.590937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.590948 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.590964 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.590974 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:01Z","lastTransitionTime":"2025-10-03T16:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.693859 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.693970 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.693982 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.693997 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.694006 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:01Z","lastTransitionTime":"2025-10-03T16:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.797607 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.797663 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.797672 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.797688 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.797699 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:01Z","lastTransitionTime":"2025-10-03T16:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.890898 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:01 crc kubenswrapper[4744]: E1003 16:26:01.891048 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.891070 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.891140 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.891086 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:01 crc kubenswrapper[4744]: E1003 16:26:01.891401 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:01 crc kubenswrapper[4744]: E1003 16:26:01.891593 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:01 crc kubenswrapper[4744]: E1003 16:26:01.891731 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.901105 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.901173 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.901194 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.901224 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:01 crc kubenswrapper[4744]: I1003 16:26:01.901247 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:01Z","lastTransitionTime":"2025-10-03T16:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.004403 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.004564 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.004626 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.004656 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.004676 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:02Z","lastTransitionTime":"2025-10-03T16:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.108738 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.108823 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.108844 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.109268 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.109558 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:02Z","lastTransitionTime":"2025-10-03T16:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.213560 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.214051 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.214201 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.214361 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.214521 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:02Z","lastTransitionTime":"2025-10-03T16:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.318403 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.318475 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.318525 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.318574 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.318601 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:02Z","lastTransitionTime":"2025-10-03T16:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.421457 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.421553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.421570 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.421607 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.421625 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:02Z","lastTransitionTime":"2025-10-03T16:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.525240 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.525302 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.525318 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.525342 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.525363 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:02Z","lastTransitionTime":"2025-10-03T16:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.629700 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.629776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.629796 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.629829 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.629851 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:02Z","lastTransitionTime":"2025-10-03T16:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.734093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.734151 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.734170 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.734197 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.734216 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:02Z","lastTransitionTime":"2025-10-03T16:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.837342 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.837413 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.837436 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.837464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.837484 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:02Z","lastTransitionTime":"2025-10-03T16:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.941251 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.941301 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.941314 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.941333 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:02 crc kubenswrapper[4744]: I1003 16:26:02.941347 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:02Z","lastTransitionTime":"2025-10-03T16:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.045191 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.045238 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.045252 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.045272 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.045285 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:03Z","lastTransitionTime":"2025-10-03T16:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.148781 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.148838 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.148851 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.148872 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.148886 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:03Z","lastTransitionTime":"2025-10-03T16:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.163565 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.163672 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.163694 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.163727 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.163747 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:03Z","lastTransitionTime":"2025-10-03T16:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:03 crc kubenswrapper[4744]: E1003 16:26:03.182736 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.188003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.188061 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.188078 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.188104 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.188166 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:03Z","lastTransitionTime":"2025-10-03T16:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:03 crc kubenswrapper[4744]: E1003 16:26:03.204214 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.209621 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.209683 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.209702 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.209729 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.209749 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:03Z","lastTransitionTime":"2025-10-03T16:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:03 crc kubenswrapper[4744]: E1003 16:26:03.227817 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.233298 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.233358 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.233381 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.233409 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.233426 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:03Z","lastTransitionTime":"2025-10-03T16:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:03 crc kubenswrapper[4744]: E1003 16:26:03.248064 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.252592 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.252650 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.252670 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.252699 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.252723 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:03Z","lastTransitionTime":"2025-10-03T16:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:03 crc kubenswrapper[4744]: E1003 16:26:03.270275 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:03Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:03 crc kubenswrapper[4744]: E1003 16:26:03.270532 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.272755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.272824 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.272838 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.272868 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.272892 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:03Z","lastTransitionTime":"2025-10-03T16:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.375650 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.375699 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.375712 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.375735 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.375748 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:03Z","lastTransitionTime":"2025-10-03T16:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.479289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.479351 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.479364 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.479387 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.479400 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:03Z","lastTransitionTime":"2025-10-03T16:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.583885 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.583959 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.583977 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.584001 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.584014 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:03Z","lastTransitionTime":"2025-10-03T16:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.687337 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.687401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.687414 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.687427 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.687439 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:03Z","lastTransitionTime":"2025-10-03T16:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.790524 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.790595 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.790612 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.790632 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.790643 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:03Z","lastTransitionTime":"2025-10-03T16:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.890979 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.891331 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.891352 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.891360 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:03 crc kubenswrapper[4744]: E1003 16:26:03.891972 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:03 crc kubenswrapper[4744]: E1003 16:26:03.892021 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:03 crc kubenswrapper[4744]: E1003 16:26:03.892124 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:03 crc kubenswrapper[4744]: E1003 16:26:03.892424 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.894464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.894541 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.894559 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.894583 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.894603 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:03Z","lastTransitionTime":"2025-10-03T16:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.914143 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.997959 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.998027 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.998045 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.998071 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:03 crc kubenswrapper[4744]: I1003 16:26:03.998090 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:03Z","lastTransitionTime":"2025-10-03T16:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.101287 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.101369 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.101400 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.101431 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.101454 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:04Z","lastTransitionTime":"2025-10-03T16:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.205129 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.205186 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.205203 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.205225 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.205238 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:04Z","lastTransitionTime":"2025-10-03T16:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.307709 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.308064 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.308202 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.308310 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.308404 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:04Z","lastTransitionTime":"2025-10-03T16:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.412462 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.412536 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.412549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.412569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.412582 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:04Z","lastTransitionTime":"2025-10-03T16:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.516433 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.516543 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.516563 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.516590 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.516606 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:04Z","lastTransitionTime":"2025-10-03T16:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.619822 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.619895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.619914 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.619947 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.619975 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:04Z","lastTransitionTime":"2025-10-03T16:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.724446 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.724550 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.724570 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.724597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.724616 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:04Z","lastTransitionTime":"2025-10-03T16:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.828682 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.828759 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.828780 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.828807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.828826 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:04Z","lastTransitionTime":"2025-10-03T16:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.910138 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d36c4bf-d76c-431a-958e-744f6e720500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475eeb458ca840cc4a5d5c82b30a55bd6f245681131113878dc984915a9dc31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a3b07aa960b742c202322c07932081b0fcc88f406dd1adc5c33562376d7b9ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a3b07aa960b742c202322c07932081b0fcc88f406dd1adc5c33562376d7b9ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.931294 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.931331 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.931340 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.931357 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.931367 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:04Z","lastTransitionTime":"2025-10-03T16:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.933787 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.947289 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.959796 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.976087 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:04 crc kubenswrapper[4744]: I1003 16:26:04.990439 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:04Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.016320 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:58Z\\\",\\\"message\\\":\\\"object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb\\\\nI1003 16:25:58.946365 6793 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1003 16:25:58.946371 6793 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-bdrsw\\\\nI1003 16:25:58.946377 6793 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF1003 16:25:58.946375 6793 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z]\\\\nI1003 16:25:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.032473 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81781145-978c-49ef-bb8e-4bf4f5eab778\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011b2e41b97e95ff1a11ff71d90fc00813ab18eac48b6cc7a987039494649a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6b1c34f972682f8afe389a96e2b35b848e4ceb02cb9379fa91c764f2cf11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52cbfb9a1c58ce525029b4226599a8e258453a8c5837643a6fd1efe2441641a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.038447 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.038515 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.038530 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.038550 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.038564 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:05Z","lastTransitionTime":"2025-10-03T16:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.048532 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.063202 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.077310 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.093534 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3888e7ce3b4e4e7967e3c7ec2e5ef8b7bee43a7742087ca6ea6f44a3d72e45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:44Z\\\",\\\"message\\\":\\\"2025-10-03T16:24:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41\\\\n2025-10-03T16:24:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41 to /host/opt/cni/bin/\\\\n2025-10-03T16:24:59Z [verbose] multus-daemon started\\\\n2025-10-03T16:24:59Z [verbose] Readiness Indicator file check\\\\n2025-10-03T16:25:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.106771 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.123184 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.136468 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.140625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.140663 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.140677 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.140697 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.140707 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:05Z","lastTransitionTime":"2025-10-03T16:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.148972 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.161955 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.176836 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.191739 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:05Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.244179 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.244235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.244244 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.244261 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.244271 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:05Z","lastTransitionTime":"2025-10-03T16:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.347150 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.347206 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.347222 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.347238 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.347253 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:05Z","lastTransitionTime":"2025-10-03T16:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.450227 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.450376 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.450397 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.450428 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.450447 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:05Z","lastTransitionTime":"2025-10-03T16:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.553232 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.553283 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.553296 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.553315 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.553329 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:05Z","lastTransitionTime":"2025-10-03T16:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.657976 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.658044 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.658064 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.658092 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.658112 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:05Z","lastTransitionTime":"2025-10-03T16:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.761478 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.761556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.761571 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.761592 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.761605 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:05Z","lastTransitionTime":"2025-10-03T16:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.864600 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.864650 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.864662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.864676 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.864686 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:05Z","lastTransitionTime":"2025-10-03T16:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.891393 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.891461 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.891531 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:05 crc kubenswrapper[4744]: E1003 16:26:05.891714 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:05 crc kubenswrapper[4744]: E1003 16:26:05.891837 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.891447 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:05 crc kubenswrapper[4744]: E1003 16:26:05.891938 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:05 crc kubenswrapper[4744]: E1003 16:26:05.892092 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.967133 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.967165 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.967173 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.967186 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:05 crc kubenswrapper[4744]: I1003 16:26:05.967198 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:05Z","lastTransitionTime":"2025-10-03T16:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.069428 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.069487 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.069519 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.069537 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.069549 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:06Z","lastTransitionTime":"2025-10-03T16:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.172006 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.172066 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.172078 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.172095 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.172111 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:06Z","lastTransitionTime":"2025-10-03T16:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.274352 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.274434 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.274463 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.274527 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.274551 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:06Z","lastTransitionTime":"2025-10-03T16:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.377661 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.377714 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.377728 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.377745 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.377778 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:06Z","lastTransitionTime":"2025-10-03T16:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.480679 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.480747 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.480764 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.480786 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.480799 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:06Z","lastTransitionTime":"2025-10-03T16:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.583049 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.583110 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.583122 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.583141 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.583156 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:06Z","lastTransitionTime":"2025-10-03T16:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.686807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.686857 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.686869 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.686886 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.686898 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:06Z","lastTransitionTime":"2025-10-03T16:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.790766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.790851 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.790874 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.790906 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.790931 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:06Z","lastTransitionTime":"2025-10-03T16:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.893279 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.893350 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.893371 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.893401 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.893423 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:06Z","lastTransitionTime":"2025-10-03T16:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.998161 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.998240 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.998267 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.998303 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:06 crc kubenswrapper[4744]: I1003 16:26:06.998329 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:06Z","lastTransitionTime":"2025-10-03T16:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.102776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.102841 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.102861 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.102888 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.102907 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:07Z","lastTransitionTime":"2025-10-03T16:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.206240 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.206318 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.206468 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.206538 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.206559 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:07Z","lastTransitionTime":"2025-10-03T16:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.309858 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.309918 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.309931 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.309956 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.309970 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:07Z","lastTransitionTime":"2025-10-03T16:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.413149 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.414224 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.414436 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.414671 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.414864 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:07Z","lastTransitionTime":"2025-10-03T16:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.518456 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.518539 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.518553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.518573 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.518589 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:07Z","lastTransitionTime":"2025-10-03T16:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.621976 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.622014 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.622023 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.622036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.622045 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:07Z","lastTransitionTime":"2025-10-03T16:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.724650 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.724717 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.724727 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.724743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.724755 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:07Z","lastTransitionTime":"2025-10-03T16:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.827759 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.828059 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.828170 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.828258 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.828338 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:07Z","lastTransitionTime":"2025-10-03T16:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.891543 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.891689 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.891741 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.891783 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:07 crc kubenswrapper[4744]: E1003 16:26:07.891935 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:07 crc kubenswrapper[4744]: E1003 16:26:07.892182 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:07 crc kubenswrapper[4744]: E1003 16:26:07.892292 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:07 crc kubenswrapper[4744]: E1003 16:26:07.892388 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.930937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.930974 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.930986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.931001 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:07 crc kubenswrapper[4744]: I1003 16:26:07.931010 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:07Z","lastTransitionTime":"2025-10-03T16:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.033420 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.033461 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.033471 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.033484 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.033514 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:08Z","lastTransitionTime":"2025-10-03T16:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.136671 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.136713 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.136721 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.136737 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.136746 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:08Z","lastTransitionTime":"2025-10-03T16:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.239735 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.239787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.239797 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.239812 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.239823 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:08Z","lastTransitionTime":"2025-10-03T16:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.343599 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.343704 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.343723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.343785 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.343807 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:08Z","lastTransitionTime":"2025-10-03T16:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.446408 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.446477 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.446527 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.446558 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.446578 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:08Z","lastTransitionTime":"2025-10-03T16:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.549485 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.549555 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.549564 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.549580 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.549592 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:08Z","lastTransitionTime":"2025-10-03T16:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.653125 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.653183 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.653197 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.653218 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.653233 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:08Z","lastTransitionTime":"2025-10-03T16:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.755542 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.755605 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.755615 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.755628 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.755656 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:08Z","lastTransitionTime":"2025-10-03T16:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.859330 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.859400 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.859418 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.859447 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.859468 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:08Z","lastTransitionTime":"2025-10-03T16:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.963638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.963725 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.963760 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.963802 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:08 crc kubenswrapper[4744]: I1003 16:26:08.963824 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:08Z","lastTransitionTime":"2025-10-03T16:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.067242 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.067327 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.067346 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.067375 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.067395 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:09Z","lastTransitionTime":"2025-10-03T16:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.182532 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.182585 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.182597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.182621 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.182638 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:09Z","lastTransitionTime":"2025-10-03T16:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.285794 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.285883 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.285942 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.285973 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.285994 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:09Z","lastTransitionTime":"2025-10-03T16:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.388713 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.388783 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.388793 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.388816 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.388834 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:09Z","lastTransitionTime":"2025-10-03T16:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.492184 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.492281 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.492295 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.492321 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.492344 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:09Z","lastTransitionTime":"2025-10-03T16:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.595033 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.595117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.595137 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.595166 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.595184 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:09Z","lastTransitionTime":"2025-10-03T16:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.698874 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.698926 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.698944 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.698966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.698987 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:09Z","lastTransitionTime":"2025-10-03T16:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.802195 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.802251 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.802273 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.802294 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.802310 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:09Z","lastTransitionTime":"2025-10-03T16:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.891654 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:09 crc kubenswrapper[4744]: E1003 16:26:09.891817 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.892457 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.892613 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.892661 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:09 crc kubenswrapper[4744]: E1003 16:26:09.893325 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:09 crc kubenswrapper[4744]: E1003 16:26:09.893428 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:09 crc kubenswrapper[4744]: E1003 16:26:09.893562 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.905194 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.905474 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.905770 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.905988 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:09 crc kubenswrapper[4744]: I1003 16:26:09.906193 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:09Z","lastTransitionTime":"2025-10-03T16:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.009468 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.009549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.009565 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.009588 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.009605 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:10Z","lastTransitionTime":"2025-10-03T16:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.112309 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.112373 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.112394 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.112424 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.112455 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:10Z","lastTransitionTime":"2025-10-03T16:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.215085 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.215941 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.216014 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.216100 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.216183 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:10Z","lastTransitionTime":"2025-10-03T16:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.320617 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.321077 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.321183 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.321279 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.321353 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:10Z","lastTransitionTime":"2025-10-03T16:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.424219 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.424278 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.424291 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.424314 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.424329 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:10Z","lastTransitionTime":"2025-10-03T16:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.526715 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.526773 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.526783 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.526799 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.526809 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:10Z","lastTransitionTime":"2025-10-03T16:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.629650 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.629708 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.629721 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.629740 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.629754 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:10Z","lastTransitionTime":"2025-10-03T16:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.732141 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.732185 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.732195 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.732210 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.732220 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:10Z","lastTransitionTime":"2025-10-03T16:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.835601 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.835653 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.835664 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.835685 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.835696 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:10Z","lastTransitionTime":"2025-10-03T16:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.939057 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.939131 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.939147 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.939168 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:10 crc kubenswrapper[4744]: I1003 16:26:10.939180 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:10Z","lastTransitionTime":"2025-10-03T16:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.041598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.041663 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.041680 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.041704 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.041722 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:11Z","lastTransitionTime":"2025-10-03T16:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.145186 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.145227 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.145235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.145248 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.145256 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:11Z","lastTransitionTime":"2025-10-03T16:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.248156 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.248202 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.248210 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.248227 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.248242 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:11Z","lastTransitionTime":"2025-10-03T16:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.351300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.351386 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.351412 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.351458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.351547 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:11Z","lastTransitionTime":"2025-10-03T16:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.455056 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.455130 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.455154 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.455188 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.455207 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:11Z","lastTransitionTime":"2025-10-03T16:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.558765 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.558815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.558830 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.558847 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.558857 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:11Z","lastTransitionTime":"2025-10-03T16:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.662015 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.662074 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.662087 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.662116 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.662135 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:11Z","lastTransitionTime":"2025-10-03T16:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.764386 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.764449 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.764459 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.764481 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.764512 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:11Z","lastTransitionTime":"2025-10-03T16:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.867331 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.867386 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.867406 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.867426 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.867441 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:11Z","lastTransitionTime":"2025-10-03T16:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.891137 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.891185 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:11 crc kubenswrapper[4744]: E1003 16:26:11.891288 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.891143 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:11 crc kubenswrapper[4744]: E1003 16:26:11.891445 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.891456 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:11 crc kubenswrapper[4744]: E1003 16:26:11.891535 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:11 crc kubenswrapper[4744]: E1003 16:26:11.891589 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.970452 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.970530 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.970543 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.970560 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:11 crc kubenswrapper[4744]: I1003 16:26:11.970575 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:11Z","lastTransitionTime":"2025-10-03T16:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.073418 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.073478 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.073488 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.073529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.073539 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:12Z","lastTransitionTime":"2025-10-03T16:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.176035 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.176087 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.176098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.176115 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.176128 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:12Z","lastTransitionTime":"2025-10-03T16:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.278691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.278725 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.278735 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.278750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.278760 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:12Z","lastTransitionTime":"2025-10-03T16:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.381898 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.381952 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.381972 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.381991 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.382007 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:12Z","lastTransitionTime":"2025-10-03T16:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.484609 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.484679 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.484690 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.484706 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.484716 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:12Z","lastTransitionTime":"2025-10-03T16:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.587225 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.587304 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.587320 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.587338 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.587350 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:12Z","lastTransitionTime":"2025-10-03T16:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.690316 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.690377 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.690392 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.690411 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.690427 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:12Z","lastTransitionTime":"2025-10-03T16:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.793388 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.793478 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.793540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.793568 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.793585 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:12Z","lastTransitionTime":"2025-10-03T16:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.892994 4744 scope.go:117] "RemoveContainer" containerID="e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb" Oct 03 16:26:12 crc kubenswrapper[4744]: E1003 16:26:12.893262 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.895347 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.895410 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.895434 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.895465 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.895488 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:12Z","lastTransitionTime":"2025-10-03T16:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.999404 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.999464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.999482 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:12 crc kubenswrapper[4744]: I1003 16:26:12.999546 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:12.999566 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:12Z","lastTransitionTime":"2025-10-03T16:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.103557 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.103619 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.103642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.103665 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.103683 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:13Z","lastTransitionTime":"2025-10-03T16:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.207869 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.207946 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.207966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.207998 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.208021 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:13Z","lastTransitionTime":"2025-10-03T16:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.311047 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.311102 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.311128 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.311152 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.311165 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:13Z","lastTransitionTime":"2025-10-03T16:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.415053 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.415139 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.415164 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.415202 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.415226 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:13Z","lastTransitionTime":"2025-10-03T16:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.519091 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.519164 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.519184 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.519213 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.519235 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:13Z","lastTransitionTime":"2025-10-03T16:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.623281 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.623807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.623835 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.623871 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.623897 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:13Z","lastTransitionTime":"2025-10-03T16:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:13 crc kubenswrapper[4744]: E1003 16:26:13.649140 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:13Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.655097 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.655157 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.655170 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.655190 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.655204 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:13Z","lastTransitionTime":"2025-10-03T16:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:13 crc kubenswrapper[4744]: E1003 16:26:13.709923 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:13Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.716534 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.716638 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.716667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.716710 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.716732 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:13Z","lastTransitionTime":"2025-10-03T16:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:13 crc kubenswrapper[4744]: E1003 16:26:13.730647 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:13Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.735036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.735083 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.735092 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.735109 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.735119 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:13Z","lastTransitionTime":"2025-10-03T16:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:13 crc kubenswrapper[4744]: E1003 16:26:13.748160 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:13Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.752975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.753016 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.753031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.753052 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.753067 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:13Z","lastTransitionTime":"2025-10-03T16:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:13 crc kubenswrapper[4744]: E1003 16:26:13.766960 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T16:26:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5c82bbbe-55e2-4b37-ab80-2af91bc4605d\\\",\\\"systemUUID\\\":\\\"1aea8ebf-25a8-4f05-acb0-bcaaf1af2496\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:13Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:13 crc kubenswrapper[4744]: E1003 16:26:13.767082 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.768846 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.768918 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.768929 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.768942 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.768951 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:13Z","lastTransitionTime":"2025-10-03T16:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.872148 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.872207 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.872221 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.872240 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.872253 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:13Z","lastTransitionTime":"2025-10-03T16:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.891798 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.891799 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:13 crc kubenswrapper[4744]: E1003 16:26:13.891994 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.891829 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.891809 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:13 crc kubenswrapper[4744]: E1003 16:26:13.892210 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:13 crc kubenswrapper[4744]: E1003 16:26:13.892112 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:13 crc kubenswrapper[4744]: E1003 16:26:13.892291 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.950714 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs\") pod \"network-metrics-daemon-4f7dk\" (UID: \"1008bf15-63d7-45f4-b42b-a4680b7dc232\") " pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:13 crc kubenswrapper[4744]: E1003 16:26:13.950949 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:26:13 crc kubenswrapper[4744]: E1003 16:26:13.951072 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs podName:1008bf15-63d7-45f4-b42b-a4680b7dc232 nodeName:}" failed. No retries permitted until 2025-10-03 16:27:17.951049953 +0000 UTC m=+164.230925849 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs") pod "network-metrics-daemon-4f7dk" (UID: "1008bf15-63d7-45f4-b42b-a4680b7dc232") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.974414 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.974461 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.974477 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.974519 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:13 crc kubenswrapper[4744]: I1003 16:26:13.974536 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:13Z","lastTransitionTime":"2025-10-03T16:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.077681 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.077737 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.077746 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.077760 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.077769 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:14Z","lastTransitionTime":"2025-10-03T16:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.180405 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.180451 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.180463 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.180480 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.180507 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:14Z","lastTransitionTime":"2025-10-03T16:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.283072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.283169 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.283187 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.283272 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.283290 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:14Z","lastTransitionTime":"2025-10-03T16:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.386120 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.386173 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.386183 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.386197 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.386209 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:14Z","lastTransitionTime":"2025-10-03T16:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.488745 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.488791 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.488803 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.488818 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.488831 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:14Z","lastTransitionTime":"2025-10-03T16:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.590936 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.590975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.590984 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.590998 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.591010 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:14Z","lastTransitionTime":"2025-10-03T16:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.693642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.693684 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.693693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.693706 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.693717 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:14Z","lastTransitionTime":"2025-10-03T16:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.795278 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.795368 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.795389 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.795421 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.795441 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:14Z","lastTransitionTime":"2025-10-03T16:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.898043 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.898099 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.898111 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.898130 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.898142 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:14Z","lastTransitionTime":"2025-10-03T16:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.921887 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ce4b9277-3689-4977-9131-7c6554fe1984\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a73cacaf974b35b18d014740c3a41aeaaa01180a6042a3018cb9cded693fa030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://551e1bbf5fe027ea67cb504a1610cdda82029409439d4d469306033c67837c64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38708903d0fb1974af9baedf606f07340c0b91b023574f0d7f1735109e8b2d19\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d57d1eeb4a6fb689d2299a56ce7a371338e7941de162ac1bd32d4ef748b52ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c8e77f0fc7fc050fea4d443a3f1b507c72e07b6d24d9a5d2a3126d0a4e35fe8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T16:24:48Z\\\",\\\"message\\\":\\\"W1003 16:24:37.950363 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1003 16:24:37.950726 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759508677 cert, and key in /tmp/serving-cert-2498967001/serving-signer.crt, /tmp/serving-cert-2498967001/serving-signer.key\\\\nI1003 16:24:38.239798 1 observer_polling.go:159] Starting file observer\\\\nW1003 16:24:38.243269 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1003 16:24:38.243450 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 16:24:38.244280 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2498967001/tls.crt::/tmp/serving-cert-2498967001/tls.key\\\\\\\"\\\\nF1003 16:24:48.726804 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e7ce25319636c0c01fcb8652ad6146979c9131185d96443678c3138f53e1509\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3abc6da65f3b2a8789cfc7a831fc16490a764d953442a44bc633d57eb0746f0b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:14Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.943987 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75f2c742-9297-4bc7-91dc-84589d660e9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a94fc2cbd90832a8aa539f82fb2a185cecde8fd24a6d272a52c934de3010eed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7632f60cb6cfe1e9e5646520feaae074c0298cc2e3a7d45f63a30a399b9dbce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43b2498efc29b9eee7ef4979cfc226e67b1557087cf6a417ddff8726b4349651\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d54fdc764b55c47e1d042e4cfbf29e7d9722e78228bfacbcb47b90f7d56c54f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:14Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.961055 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:14Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.976368 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-pmktb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"134d85a0-99b6-4e3d-8960-5b168792a0ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0dfbe01ef5d21487de8c6ec4778fa20367b300872baeab54510e8022cf07cfe9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-blqh6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-pmktb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:14Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:14 crc kubenswrapper[4744]: I1003 16:26:14.989839 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6rjqm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e799ae45-a110-4d69-8c66-3d89111eab93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccb2e5f92c99317ead27a40ea77d9c6f292e14083c74ae8ad021cb230baccc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pd2s8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6rjqm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:14Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.006707 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc860549-945d-4922-a258-99170f6912c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9941a2138a23555206a001276c99410317a61c550f0bebcb76bab26fdf080963\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f073483247c659b448e4a69c021bdbcdcea48643664600a0dfcb60792311aaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-45hkm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-425zb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.009680 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.009759 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.009791 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.009819 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.009834 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:15Z","lastTransitionTime":"2025-10-03T16:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.027983 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1008bf15-63d7-45f4-b42b-a4680b7dc232\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zkz2k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:25:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4f7dk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.038407 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d36c4bf-d76c-431a-958e-744f6e720500\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a475eeb458ca840cc4a5d5c82b30a55bd6f245681131113878dc984915a9dc31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a3b07aa960b742c202322c07932081b0fcc88f406dd1adc5c33562376d7b9ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a3b07aa960b742c202322c07932081b0fcc88f406dd1adc5c33562376d7b9ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.056303 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9e3cbc-a10e-443b-97aa-73015e18d37a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4db1c163ed0214977577725243aadb74123948d483ff5a339b6952af77456a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fc6eb578160a682a3a52d9e2d202e5cc78a22f26249a8a1c24f8b84db5cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e5622e429a7fae91b87d5632a88f82588a5fe6a9fd00a7e37603027e11454a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff701e7154af7d34310eb47cb99b474709c6b48a084f80491919c32ef7515d4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d20e963a5774fd9f6945201c42f7914cb465036a3b8046c7079b60818c77134\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92cc800bcde63f67df00e226b19c23ece62a3cfef4b7b6a81ae1e34540ffd458\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0208bb04a5865e2ff7f64e0b0b3a97ede3a6414a7ad111a3aed1fd215dc560c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07dac168d12ba0cc2b6c6006cbe65690932c6ca771a77a6a07c14ee34f230319\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.068159 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35caa272d85751fda1567d5c0c0658e20eae71f4ee192f9f1abc44682a7d879e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bf4214c578367c688cfd47ed2ef014f6c0ff02a84d64af7af06ccb1d87ea95e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.079003 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2f773103a0a4d8c99b4035e5daefb21d5106494b26fd1b669f9401c651e3d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.093153 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6190db0e-c421-41e5-acc2-cf96c34fe3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f13e009e2d798c4b3131e12f2f10473feb7e870167d7917a091c1b5d5674a44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ebd780700c5d0685cadcc940fdafb990e12291f6144e0b72da447edde0d47b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f991f0a157afd2bb122b577fe934c2b99df6bb8b0477d41b2946b04174249505\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6355e154210676dc272e1ef4d227db08e86669b069b1b8d065f84946d15d1037\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://783f1508f55cf5e92537ebc26c4b89f467eb8a60b39f2b98a7a2dfdba48aa02a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b334e2b2fc67424b2a0e8d202b109d923c76c4acff191afbf2bfdd8f346a82c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d564699e790216003fbd5d7fca0c7d99cf41358a1619a5b0f5d83934eef3583\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:25:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-knqt4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-bdrsw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.106225 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cc9bea96-4847-44c8-ab50-d469a6c6155d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dcabdc8890d995f3a49d725e96e2fc91138499fcf2a5e3681e0715149c1083d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rqp6h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qzw8f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.113481 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.113575 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.113589 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.113612 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.113626 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:15Z","lastTransitionTime":"2025-10-03T16:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.125116 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3436425f-b9cf-4302-b5ad-ccb92d597aed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:58Z\\\",\\\"message\\\":\\\"object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb\\\\nI1003 16:25:58.946365 6793 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1003 16:25:58.946371 6793 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-bdrsw\\\\nI1003 16:25:58.946377 6793 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF1003 16:25:58.946375 6793 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:25:58Z is after 2025-08-24T17:21:41Z]\\\\nI1003 16:25:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:25:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nxw7s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-jw5rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.137744 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81781145-978c-49ef-bb8e-4bf4f5eab778\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011b2e41b97e95ff1a11ff71d90fc00813ab18eac48b6cc7a987039494649a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e6b1c34f972682f8afe389a96e2b35b848e4ceb02cb9379fa91c764f2cf11e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52cbfb9a1c58ce525029b4226599a8e258453a8c5837643a6fd1efe2441641a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3de322380be12d46e21f38d4ce2ca793957f01b83242f08a6f63ca1fe24a3e7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T16:24:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.149690 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.166404 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e2bdb6739c4327bdec7fced030375cd44bb9128385de651f04b470934a8f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.187019 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:53Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.200051 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vqz6q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84192d79-366a-453f-b70b-aefa4537ec4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:24:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T16:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3888e7ce3b4e4e7967e3c7ec2e5ef8b7bee43a7742087ca6ea6f44a3d72e45c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T16:25:44Z\\\",\\\"message\\\":\\\"2025-10-03T16:24:58+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41\\\\n2025-10-03T16:24:58+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_56490e0f-c07a-4671-a56d-968a0d760a41 to /host/opt/cni/bin/\\\\n2025-10-03T16:24:59Z [verbose] multus-daemon started\\\\n2025-10-03T16:24:59Z [verbose] Readiness Indicator file check\\\\n2025-10-03T16:25:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T16:24:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T16:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5b757\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T16:24:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vqz6q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T16:26:15Z is after 2025-08-24T17:21:41Z" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.216053 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.216108 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.216124 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.216146 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.216159 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:15Z","lastTransitionTime":"2025-10-03T16:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.319179 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.319252 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.319274 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.319305 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.319325 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:15Z","lastTransitionTime":"2025-10-03T16:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.422874 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.422946 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.422965 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.422990 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.423008 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:15Z","lastTransitionTime":"2025-10-03T16:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.525944 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.526328 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.526393 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.526460 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.526546 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:15Z","lastTransitionTime":"2025-10-03T16:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.629728 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.629775 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.629787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.629810 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.629824 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:15Z","lastTransitionTime":"2025-10-03T16:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.733970 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.734035 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.734049 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.734075 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.734093 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:15Z","lastTransitionTime":"2025-10-03T16:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.837672 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.837717 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.837729 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.837750 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.837762 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:15Z","lastTransitionTime":"2025-10-03T16:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.891727 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.891774 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.891774 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.891900 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:15 crc kubenswrapper[4744]: E1003 16:26:15.891972 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:15 crc kubenswrapper[4744]: E1003 16:26:15.892079 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:15 crc kubenswrapper[4744]: E1003 16:26:15.892267 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:15 crc kubenswrapper[4744]: E1003 16:26:15.892459 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.940459 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.940594 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.940628 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.940668 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:15 crc kubenswrapper[4744]: I1003 16:26:15.940703 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:15Z","lastTransitionTime":"2025-10-03T16:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.044001 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.044383 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.044457 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.044545 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.044641 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:16Z","lastTransitionTime":"2025-10-03T16:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.147653 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.147988 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.148169 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.148279 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.148447 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:16Z","lastTransitionTime":"2025-10-03T16:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.250819 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.251134 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.251304 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.251418 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.251523 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:16Z","lastTransitionTime":"2025-10-03T16:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.354748 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.354826 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.354848 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.354878 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.354898 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:16Z","lastTransitionTime":"2025-10-03T16:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.458214 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.458277 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.458295 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.458319 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.458334 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:16Z","lastTransitionTime":"2025-10-03T16:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.562004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.562059 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.562075 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.562095 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.562109 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:16Z","lastTransitionTime":"2025-10-03T16:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.666121 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.666186 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.666204 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.666230 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.666252 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:16Z","lastTransitionTime":"2025-10-03T16:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.769691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.769760 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.769787 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.769820 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.769845 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:16Z","lastTransitionTime":"2025-10-03T16:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.873725 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.873799 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.873821 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.873849 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.873869 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:16Z","lastTransitionTime":"2025-10-03T16:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.977292 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.977360 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.977387 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.977422 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:16 crc kubenswrapper[4744]: I1003 16:26:16.977449 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:16Z","lastTransitionTime":"2025-10-03T16:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.081307 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.081371 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.081382 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.081405 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.081417 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:17Z","lastTransitionTime":"2025-10-03T16:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.185633 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.185709 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.185734 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.185768 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.185794 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:17Z","lastTransitionTime":"2025-10-03T16:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.290083 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.290621 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.290797 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.291005 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.291196 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:17Z","lastTransitionTime":"2025-10-03T16:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.396001 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.396084 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.396112 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.396146 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.396170 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:17Z","lastTransitionTime":"2025-10-03T16:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.500327 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.500403 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.500425 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.500458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.500485 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:17Z","lastTransitionTime":"2025-10-03T16:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.603621 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.603681 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.603694 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.603717 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.603735 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:17Z","lastTransitionTime":"2025-10-03T16:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.706799 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.706847 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.706859 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.706881 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.706895 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:17Z","lastTransitionTime":"2025-10-03T16:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.810737 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.810814 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.810835 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.810864 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.810884 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:17Z","lastTransitionTime":"2025-10-03T16:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.891874 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.891960 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.892194 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:17 crc kubenswrapper[4744]: E1003 16:26:17.892334 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.892441 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:17 crc kubenswrapper[4744]: E1003 16:26:17.892543 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:17 crc kubenswrapper[4744]: E1003 16:26:17.892647 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:17 crc kubenswrapper[4744]: E1003 16:26:17.892732 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.914160 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.914192 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.914203 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.914221 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:17 crc kubenswrapper[4744]: I1003 16:26:17.914234 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:17Z","lastTransitionTime":"2025-10-03T16:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.017808 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.018300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.018539 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.018741 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.018962 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:18Z","lastTransitionTime":"2025-10-03T16:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.122097 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.122161 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.122178 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.122202 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.122216 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:18Z","lastTransitionTime":"2025-10-03T16:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.225975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.226045 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.226064 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.226095 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.226119 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:18Z","lastTransitionTime":"2025-10-03T16:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.330250 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.330322 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.330345 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.330381 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.330408 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:18Z","lastTransitionTime":"2025-10-03T16:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.434031 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.434108 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.434129 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.434158 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.434180 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:18Z","lastTransitionTime":"2025-10-03T16:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.538041 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.538093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.538124 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.538171 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.538197 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:18Z","lastTransitionTime":"2025-10-03T16:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.641971 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.642033 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.642057 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.642088 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.642110 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:18Z","lastTransitionTime":"2025-10-03T16:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.745052 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.745136 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.745160 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.745193 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.745227 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:18Z","lastTransitionTime":"2025-10-03T16:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.848534 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.848604 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.848622 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.848654 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.848674 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:18Z","lastTransitionTime":"2025-10-03T16:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.951484 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.951601 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.951617 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.951642 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:18 crc kubenswrapper[4744]: I1003 16:26:18.951659 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:18Z","lastTransitionTime":"2025-10-03T16:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.055645 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.055720 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.055740 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.055783 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.055818 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:19Z","lastTransitionTime":"2025-10-03T16:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.160135 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.160198 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.160215 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.160240 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.160260 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:19Z","lastTransitionTime":"2025-10-03T16:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.264008 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.264081 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.264098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.264120 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.264134 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:19Z","lastTransitionTime":"2025-10-03T16:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.367289 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.367366 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.367385 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.367419 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.367441 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:19Z","lastTransitionTime":"2025-10-03T16:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.469778 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.469831 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.469840 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.469860 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.469871 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:19Z","lastTransitionTime":"2025-10-03T16:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.573177 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.573261 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.573281 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.573306 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.573325 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:19Z","lastTransitionTime":"2025-10-03T16:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.676089 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.676155 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.676175 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.676202 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.676225 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:19Z","lastTransitionTime":"2025-10-03T16:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.780201 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.780254 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.780272 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.780299 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.780320 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:19Z","lastTransitionTime":"2025-10-03T16:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.883717 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.883788 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.883802 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.883825 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.883843 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:19Z","lastTransitionTime":"2025-10-03T16:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.890951 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.890964 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.891074 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:19 crc kubenswrapper[4744]: E1003 16:26:19.891139 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.891160 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:19 crc kubenswrapper[4744]: E1003 16:26:19.891301 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:19 crc kubenswrapper[4744]: E1003 16:26:19.891380 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:19 crc kubenswrapper[4744]: E1003 16:26:19.891621 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.988362 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.988423 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.988439 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.988467 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:19 crc kubenswrapper[4744]: I1003 16:26:19.988486 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:19Z","lastTransitionTime":"2025-10-03T16:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.093225 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.093288 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.093304 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.093324 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.093336 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:20Z","lastTransitionTime":"2025-10-03T16:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.197241 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.197326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.197341 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.197391 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.197407 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:20Z","lastTransitionTime":"2025-10-03T16:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.300824 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.300882 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.300894 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.300914 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.300928 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:20Z","lastTransitionTime":"2025-10-03T16:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.404138 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.404189 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.404201 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.404244 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.404259 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:20Z","lastTransitionTime":"2025-10-03T16:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.507148 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.507200 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.507212 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.507229 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.507242 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:20Z","lastTransitionTime":"2025-10-03T16:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.610209 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.610273 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.610291 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.610308 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.610320 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:20Z","lastTransitionTime":"2025-10-03T16:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.714065 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.714114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.714128 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.714148 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.714164 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:20Z","lastTransitionTime":"2025-10-03T16:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.816309 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.816361 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.816374 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.816389 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.816397 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:20Z","lastTransitionTime":"2025-10-03T16:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.919789 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.919894 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.919914 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.919939 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:20 crc kubenswrapper[4744]: I1003 16:26:20.919957 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:20Z","lastTransitionTime":"2025-10-03T16:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.023127 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.023230 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.023247 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.023305 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.023335 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:21Z","lastTransitionTime":"2025-10-03T16:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.126427 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.126507 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.126518 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.126541 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.126550 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:21Z","lastTransitionTime":"2025-10-03T16:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.229809 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.229894 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.229910 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.229930 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.229944 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:21Z","lastTransitionTime":"2025-10-03T16:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.333006 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.333061 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.333072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.333088 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.333100 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:21Z","lastTransitionTime":"2025-10-03T16:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.436133 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.436209 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.436229 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.436261 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.436282 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:21Z","lastTransitionTime":"2025-10-03T16:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.539428 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.539567 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.539582 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.539601 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.539617 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:21Z","lastTransitionTime":"2025-10-03T16:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.642937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.642994 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.643005 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.643022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.643033 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:21Z","lastTransitionTime":"2025-10-03T16:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.744900 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.744946 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.744957 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.744971 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.744983 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:21Z","lastTransitionTime":"2025-10-03T16:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.847279 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.847334 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.847343 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.847360 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.847375 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:21Z","lastTransitionTime":"2025-10-03T16:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.891218 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.891277 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.891231 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:21 crc kubenswrapper[4744]: E1003 16:26:21.891476 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:21 crc kubenswrapper[4744]: E1003 16:26:21.891848 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:21 crc kubenswrapper[4744]: E1003 16:26:21.891976 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.892138 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:21 crc kubenswrapper[4744]: E1003 16:26:21.892594 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.950038 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.950075 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.950084 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.950098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:21 crc kubenswrapper[4744]: I1003 16:26:21.950108 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:21Z","lastTransitionTime":"2025-10-03T16:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.052037 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.052065 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.052072 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.052086 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.052094 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:22Z","lastTransitionTime":"2025-10-03T16:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.154600 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.154646 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.154658 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.154673 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.154685 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:22Z","lastTransitionTime":"2025-10-03T16:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.257631 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.257680 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.257695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.257714 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.257726 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:22Z","lastTransitionTime":"2025-10-03T16:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.360695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.360756 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.360766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.360783 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.360796 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:22Z","lastTransitionTime":"2025-10-03T16:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.462953 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.463010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.463023 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.463041 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.463054 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:22Z","lastTransitionTime":"2025-10-03T16:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.565410 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.565452 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.565461 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.565477 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.565486 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:22Z","lastTransitionTime":"2025-10-03T16:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.667540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.667610 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.667623 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.667643 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.667653 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:22Z","lastTransitionTime":"2025-10-03T16:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.770989 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.771045 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.771063 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.771086 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.771104 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:22Z","lastTransitionTime":"2025-10-03T16:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.873119 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.873186 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.873206 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.873231 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.873247 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:22Z","lastTransitionTime":"2025-10-03T16:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.976211 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.976258 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.976300 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.976314 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:22 crc kubenswrapper[4744]: I1003 16:26:22.976324 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:22Z","lastTransitionTime":"2025-10-03T16:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.078918 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.078970 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.078985 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.079003 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.079018 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:23Z","lastTransitionTime":"2025-10-03T16:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.182036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.182093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.182102 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.182133 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.182152 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:23Z","lastTransitionTime":"2025-10-03T16:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.284842 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.284885 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.284895 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.284912 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.284922 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:23Z","lastTransitionTime":"2025-10-03T16:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.387794 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.387854 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.387866 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.387884 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.387896 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:23Z","lastTransitionTime":"2025-10-03T16:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.489973 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.490038 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.490051 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.490070 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.490085 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:23Z","lastTransitionTime":"2025-10-03T16:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.593261 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.593326 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.593340 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.593364 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.593377 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:23Z","lastTransitionTime":"2025-10-03T16:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.696337 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.696408 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.696428 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.696461 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.696523 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:23Z","lastTransitionTime":"2025-10-03T16:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.793528 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.793733 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.793774 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.793816 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.793843 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:23Z","lastTransitionTime":"2025-10-03T16:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.832385 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.832444 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.832454 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.832474 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.832493 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T16:26:23Z","lastTransitionTime":"2025-10-03T16:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.861919 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj"] Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.862355 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.864638 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.865002 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.865061 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.868068 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.891801 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:23 crc kubenswrapper[4744]: E1003 16:26:23.891979 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.892044 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:23 crc kubenswrapper[4744]: E1003 16:26:23.892114 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.892154 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:23 crc kubenswrapper[4744]: E1003 16:26:23.892194 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.892226 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:23 crc kubenswrapper[4744]: E1003 16:26:23.892264 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.892968 4744 scope.go:117] "RemoveContainer" containerID="e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb" Oct 03 16:26:23 crc kubenswrapper[4744]: E1003 16:26:23.893130 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.900438 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-vqz6q" podStartSLOduration=88.900411798 podStartE2EDuration="1m28.900411798s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:26:23.899994208 +0000 UTC m=+110.179870114" watchObservedRunningTime="2025-10-03 16:26:23.900411798 +0000 UTC m=+110.180287714" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.919541 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podStartSLOduration=88.919479 podStartE2EDuration="1m28.919479s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:26:23.919079941 +0000 UTC m=+110.198955837" watchObservedRunningTime="2025-10-03 16:26:23.919479 +0000 UTC m=+110.199354916" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.964322 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=60.964302964 podStartE2EDuration="1m0.964302964s" podCreationTimestamp="2025-10-03 16:25:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:26:23.963799703 +0000 UTC m=+110.243675599" watchObservedRunningTime="2025-10-03 16:26:23.964302964 +0000 UTC m=+110.244178860" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.972919 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c0f982b0-5021-4f8d-bc71-610e56a8d7d0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dz9fj\" (UID: \"c0f982b0-5021-4f8d-bc71-610e56a8d7d0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.973026 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c0f982b0-5021-4f8d-bc71-610e56a8d7d0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dz9fj\" (UID: \"c0f982b0-5021-4f8d-bc71-610e56a8d7d0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.973105 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0f982b0-5021-4f8d-bc71-610e56a8d7d0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dz9fj\" (UID: \"c0f982b0-5021-4f8d-bc71-610e56a8d7d0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.973129 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c0f982b0-5021-4f8d-bc71-610e56a8d7d0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dz9fj\" (UID: \"c0f982b0-5021-4f8d-bc71-610e56a8d7d0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:23 crc kubenswrapper[4744]: I1003 16:26:23.973155 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c0f982b0-5021-4f8d-bc71-610e56a8d7d0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dz9fj\" (UID: \"c0f982b0-5021-4f8d-bc71-610e56a8d7d0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.024741 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-6rjqm" podStartSLOduration=89.024716309 podStartE2EDuration="1m29.024716309s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:26:24.024609516 +0000 UTC m=+110.304485412" watchObservedRunningTime="2025-10-03 16:26:24.024716309 +0000 UTC m=+110.304592205" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.025107 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-pmktb" podStartSLOduration=90.025102608 podStartE2EDuration="1m30.025102608s" podCreationTimestamp="2025-10-03 16:24:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:26:24.014389594 +0000 UTC m=+110.294265490" watchObservedRunningTime="2025-10-03 16:26:24.025102608 +0000 UTC m=+110.304978504" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.044468 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=91.044443887 podStartE2EDuration="1m31.044443887s" podCreationTimestamp="2025-10-03 16:24:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:26:24.04414285 +0000 UTC m=+110.324018756" watchObservedRunningTime="2025-10-03 16:26:24.044443887 +0000 UTC m=+110.324319783" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.062383 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=86.062356232 podStartE2EDuration="1m26.062356232s" podCreationTimestamp="2025-10-03 16:24:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:26:24.061898762 +0000 UTC m=+110.341774668" watchObservedRunningTime="2025-10-03 16:26:24.062356232 +0000 UTC m=+110.342232128" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.074444 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c0f982b0-5021-4f8d-bc71-610e56a8d7d0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dz9fj\" (UID: \"c0f982b0-5021-4f8d-bc71-610e56a8d7d0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.074491 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0f982b0-5021-4f8d-bc71-610e56a8d7d0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dz9fj\" (UID: \"c0f982b0-5021-4f8d-bc71-610e56a8d7d0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.074541 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c0f982b0-5021-4f8d-bc71-610e56a8d7d0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dz9fj\" (UID: \"c0f982b0-5021-4f8d-bc71-610e56a8d7d0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.074589 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c0f982b0-5021-4f8d-bc71-610e56a8d7d0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dz9fj\" (UID: \"c0f982b0-5021-4f8d-bc71-610e56a8d7d0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.074619 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c0f982b0-5021-4f8d-bc71-610e56a8d7d0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dz9fj\" (UID: \"c0f982b0-5021-4f8d-bc71-610e56a8d7d0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.074709 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c0f982b0-5021-4f8d-bc71-610e56a8d7d0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-dz9fj\" (UID: \"c0f982b0-5021-4f8d-bc71-610e56a8d7d0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.074767 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c0f982b0-5021-4f8d-bc71-610e56a8d7d0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-dz9fj\" (UID: \"c0f982b0-5021-4f8d-bc71-610e56a8d7d0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.075911 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c0f982b0-5021-4f8d-bc71-610e56a8d7d0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-dz9fj\" (UID: \"c0f982b0-5021-4f8d-bc71-610e56a8d7d0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.082156 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0f982b0-5021-4f8d-bc71-610e56a8d7d0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-dz9fj\" (UID: \"c0f982b0-5021-4f8d-bc71-610e56a8d7d0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.091711 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c0f982b0-5021-4f8d-bc71-610e56a8d7d0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-dz9fj\" (UID: \"c0f982b0-5021-4f8d-bc71-610e56a8d7d0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.096218 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-425zb" podStartSLOduration=88.096187896 podStartE2EDuration="1m28.096187896s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:26:24.09554303 +0000 UTC m=+110.375418936" watchObservedRunningTime="2025-10-03 16:26:24.096187896 +0000 UTC m=+110.376063792" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.137680 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-bdrsw" podStartSLOduration=89.13765549 podStartE2EDuration="1m29.13765549s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:26:24.137559178 +0000 UTC m=+110.417435084" watchObservedRunningTime="2025-10-03 16:26:24.13765549 +0000 UTC m=+110.417531386" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.187092 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.194430 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=87.194414618 podStartE2EDuration="1m27.194414618s" podCreationTimestamp="2025-10-03 16:24:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:26:24.194313305 +0000 UTC m=+110.474189211" watchObservedRunningTime="2025-10-03 16:26:24.194414618 +0000 UTC m=+110.474290514" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.195716 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=21.195708778 podStartE2EDuration="21.195708778s" podCreationTimestamp="2025-10-03 16:26:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:26:24.160134324 +0000 UTC m=+110.440010220" watchObservedRunningTime="2025-10-03 16:26:24.195708778 +0000 UTC m=+110.475584674" Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.493027 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" event={"ID":"c0f982b0-5021-4f8d-bc71-610e56a8d7d0","Type":"ContainerStarted","Data":"aa8873d51bbf59e9939631570ef65ccf0bb9c0d48e60108fbbbc72e0f85c9ac6"} Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.493094 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" event={"ID":"c0f982b0-5021-4f8d-bc71-610e56a8d7d0","Type":"ContainerStarted","Data":"1dee8b74f92a34da855fcb8678b9ed5d3dceb6b6709d51eccf12656e369a8198"} Oct 03 16:26:24 crc kubenswrapper[4744]: I1003 16:26:24.507886 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-dz9fj" podStartSLOduration=89.507860359 podStartE2EDuration="1m29.507860359s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:26:24.506664751 +0000 UTC m=+110.786540647" watchObservedRunningTime="2025-10-03 16:26:24.507860359 +0000 UTC m=+110.787736255" Oct 03 16:26:25 crc kubenswrapper[4744]: I1003 16:26:25.891572 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:25 crc kubenswrapper[4744]: I1003 16:26:25.891620 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:25 crc kubenswrapper[4744]: I1003 16:26:25.891588 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:25 crc kubenswrapper[4744]: E1003 16:26:25.891770 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:25 crc kubenswrapper[4744]: E1003 16:26:25.891708 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:25 crc kubenswrapper[4744]: I1003 16:26:25.891812 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:25 crc kubenswrapper[4744]: E1003 16:26:25.892082 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:25 crc kubenswrapper[4744]: E1003 16:26:25.892131 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:27 crc kubenswrapper[4744]: I1003 16:26:27.891356 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:27 crc kubenswrapper[4744]: E1003 16:26:27.891514 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:27 crc kubenswrapper[4744]: I1003 16:26:27.891485 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:27 crc kubenswrapper[4744]: I1003 16:26:27.891560 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:27 crc kubenswrapper[4744]: I1003 16:26:27.891588 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:27 crc kubenswrapper[4744]: E1003 16:26:27.891610 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:27 crc kubenswrapper[4744]: E1003 16:26:27.891778 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:27 crc kubenswrapper[4744]: E1003 16:26:27.891998 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:29 crc kubenswrapper[4744]: I1003 16:26:29.891452 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:29 crc kubenswrapper[4744]: I1003 16:26:29.891541 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:29 crc kubenswrapper[4744]: I1003 16:26:29.891639 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:29 crc kubenswrapper[4744]: E1003 16:26:29.892922 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:29 crc kubenswrapper[4744]: I1003 16:26:29.891696 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:29 crc kubenswrapper[4744]: E1003 16:26:29.892982 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:29 crc kubenswrapper[4744]: E1003 16:26:29.893130 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:29 crc kubenswrapper[4744]: E1003 16:26:29.893320 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:30 crc kubenswrapper[4744]: I1003 16:26:30.526791 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vqz6q_84192d79-366a-453f-b70b-aefa4537ec4c/kube-multus/1.log" Oct 03 16:26:30 crc kubenswrapper[4744]: I1003 16:26:30.527687 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vqz6q_84192d79-366a-453f-b70b-aefa4537ec4c/kube-multus/0.log" Oct 03 16:26:30 crc kubenswrapper[4744]: I1003 16:26:30.527792 4744 generic.go:334] "Generic (PLEG): container finished" podID="84192d79-366a-453f-b70b-aefa4537ec4c" containerID="d3888e7ce3b4e4e7967e3c7ec2e5ef8b7bee43a7742087ca6ea6f44a3d72e45c" exitCode=1 Oct 03 16:26:30 crc kubenswrapper[4744]: I1003 16:26:30.527921 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vqz6q" event={"ID":"84192d79-366a-453f-b70b-aefa4537ec4c","Type":"ContainerDied","Data":"d3888e7ce3b4e4e7967e3c7ec2e5ef8b7bee43a7742087ca6ea6f44a3d72e45c"} Oct 03 16:26:30 crc kubenswrapper[4744]: I1003 16:26:30.528038 4744 scope.go:117] "RemoveContainer" containerID="c139f91eeed4c4bdac017cae865a310312c7c5a3f6dadbe00b52cf3041b47c17" Oct 03 16:26:30 crc kubenswrapper[4744]: I1003 16:26:30.528738 4744 scope.go:117] "RemoveContainer" containerID="d3888e7ce3b4e4e7967e3c7ec2e5ef8b7bee43a7742087ca6ea6f44a3d72e45c" Oct 03 16:26:30 crc kubenswrapper[4744]: E1003 16:26:30.529057 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-vqz6q_openshift-multus(84192d79-366a-453f-b70b-aefa4537ec4c)\"" pod="openshift-multus/multus-vqz6q" podUID="84192d79-366a-453f-b70b-aefa4537ec4c" Oct 03 16:26:31 crc kubenswrapper[4744]: I1003 16:26:31.535905 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vqz6q_84192d79-366a-453f-b70b-aefa4537ec4c/kube-multus/1.log" Oct 03 16:26:31 crc kubenswrapper[4744]: I1003 16:26:31.891278 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:31 crc kubenswrapper[4744]: I1003 16:26:31.891380 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:31 crc kubenswrapper[4744]: I1003 16:26:31.891381 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:31 crc kubenswrapper[4744]: I1003 16:26:31.891302 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:31 crc kubenswrapper[4744]: E1003 16:26:31.891524 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:31 crc kubenswrapper[4744]: E1003 16:26:31.891705 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:31 crc kubenswrapper[4744]: E1003 16:26:31.891873 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:31 crc kubenswrapper[4744]: E1003 16:26:31.891936 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:33 crc kubenswrapper[4744]: I1003 16:26:33.891153 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:33 crc kubenswrapper[4744]: I1003 16:26:33.891308 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:33 crc kubenswrapper[4744]: I1003 16:26:33.891464 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:33 crc kubenswrapper[4744]: E1003 16:26:33.891456 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:33 crc kubenswrapper[4744]: E1003 16:26:33.891752 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:33 crc kubenswrapper[4744]: E1003 16:26:33.891819 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:33 crc kubenswrapper[4744]: I1003 16:26:33.891859 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:33 crc kubenswrapper[4744]: E1003 16:26:33.891943 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:34 crc kubenswrapper[4744]: E1003 16:26:34.913010 4744 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 03 16:26:35 crc kubenswrapper[4744]: E1003 16:26:35.002975 4744 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 03 16:26:35 crc kubenswrapper[4744]: I1003 16:26:35.890974 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:35 crc kubenswrapper[4744]: I1003 16:26:35.891096 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:35 crc kubenswrapper[4744]: E1003 16:26:35.891168 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:35 crc kubenswrapper[4744]: I1003 16:26:35.891167 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:35 crc kubenswrapper[4744]: E1003 16:26:35.891257 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:35 crc kubenswrapper[4744]: E1003 16:26:35.891491 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:35 crc kubenswrapper[4744]: I1003 16:26:35.891554 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:35 crc kubenswrapper[4744]: E1003 16:26:35.891757 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:36 crc kubenswrapper[4744]: I1003 16:26:36.892943 4744 scope.go:117] "RemoveContainer" containerID="e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb" Oct 03 16:26:36 crc kubenswrapper[4744]: E1003 16:26:36.893182 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-jw5rc_openshift-ovn-kubernetes(3436425f-b9cf-4302-b5ad-ccb92d597aed)\"" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" Oct 03 16:26:37 crc kubenswrapper[4744]: I1003 16:26:37.890958 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:37 crc kubenswrapper[4744]: I1003 16:26:37.890996 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:37 crc kubenswrapper[4744]: I1003 16:26:37.891022 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:37 crc kubenswrapper[4744]: I1003 16:26:37.890994 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:37 crc kubenswrapper[4744]: E1003 16:26:37.891122 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:37 crc kubenswrapper[4744]: E1003 16:26:37.891192 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:37 crc kubenswrapper[4744]: E1003 16:26:37.891321 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:37 crc kubenswrapper[4744]: E1003 16:26:37.891444 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:39 crc kubenswrapper[4744]: I1003 16:26:39.891155 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:39 crc kubenswrapper[4744]: I1003 16:26:39.891195 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:39 crc kubenswrapper[4744]: I1003 16:26:39.891213 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:39 crc kubenswrapper[4744]: I1003 16:26:39.891178 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:39 crc kubenswrapper[4744]: E1003 16:26:39.891305 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:39 crc kubenswrapper[4744]: E1003 16:26:39.891485 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:39 crc kubenswrapper[4744]: E1003 16:26:39.891534 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:39 crc kubenswrapper[4744]: E1003 16:26:39.891592 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:40 crc kubenswrapper[4744]: E1003 16:26:40.005092 4744 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 03 16:26:41 crc kubenswrapper[4744]: I1003 16:26:41.891471 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:41 crc kubenswrapper[4744]: I1003 16:26:41.891480 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:41 crc kubenswrapper[4744]: E1003 16:26:41.892267 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:41 crc kubenswrapper[4744]: E1003 16:26:41.892297 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:41 crc kubenswrapper[4744]: I1003 16:26:41.891616 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:41 crc kubenswrapper[4744]: E1003 16:26:41.892449 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:41 crc kubenswrapper[4744]: I1003 16:26:41.891657 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:41 crc kubenswrapper[4744]: E1003 16:26:41.892674 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:43 crc kubenswrapper[4744]: I1003 16:26:43.891587 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:43 crc kubenswrapper[4744]: I1003 16:26:43.891627 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:43 crc kubenswrapper[4744]: I1003 16:26:43.891650 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:43 crc kubenswrapper[4744]: I1003 16:26:43.891587 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:43 crc kubenswrapper[4744]: E1003 16:26:43.891795 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:43 crc kubenswrapper[4744]: E1003 16:26:43.892007 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:43 crc kubenswrapper[4744]: E1003 16:26:43.892204 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:43 crc kubenswrapper[4744]: E1003 16:26:43.892318 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:43 crc kubenswrapper[4744]: I1003 16:26:43.892520 4744 scope.go:117] "RemoveContainer" containerID="d3888e7ce3b4e4e7967e3c7ec2e5ef8b7bee43a7742087ca6ea6f44a3d72e45c" Oct 03 16:26:44 crc kubenswrapper[4744]: I1003 16:26:44.586156 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vqz6q_84192d79-366a-453f-b70b-aefa4537ec4c/kube-multus/1.log" Oct 03 16:26:44 crc kubenswrapper[4744]: I1003 16:26:44.586228 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vqz6q" event={"ID":"84192d79-366a-453f-b70b-aefa4537ec4c","Type":"ContainerStarted","Data":"b9031bf515dd80acd90501c31a1f575672257c7875c2e6b608cf36a4e51744a4"} Oct 03 16:26:45 crc kubenswrapper[4744]: E1003 16:26:45.006025 4744 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 03 16:26:45 crc kubenswrapper[4744]: I1003 16:26:45.891032 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:45 crc kubenswrapper[4744]: I1003 16:26:45.891078 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:45 crc kubenswrapper[4744]: I1003 16:26:45.891170 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:45 crc kubenswrapper[4744]: I1003 16:26:45.891056 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:45 crc kubenswrapper[4744]: E1003 16:26:45.891310 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:45 crc kubenswrapper[4744]: E1003 16:26:45.891691 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:45 crc kubenswrapper[4744]: E1003 16:26:45.891866 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:45 crc kubenswrapper[4744]: E1003 16:26:45.891988 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:47 crc kubenswrapper[4744]: I1003 16:26:47.891290 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:47 crc kubenswrapper[4744]: I1003 16:26:47.891375 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:47 crc kubenswrapper[4744]: I1003 16:26:47.891416 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:47 crc kubenswrapper[4744]: E1003 16:26:47.891550 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:47 crc kubenswrapper[4744]: I1003 16:26:47.891586 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:47 crc kubenswrapper[4744]: E1003 16:26:47.891823 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:47 crc kubenswrapper[4744]: E1003 16:26:47.891860 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:47 crc kubenswrapper[4744]: E1003 16:26:47.892125 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:49 crc kubenswrapper[4744]: I1003 16:26:49.891910 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:49 crc kubenswrapper[4744]: I1003 16:26:49.891910 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:49 crc kubenswrapper[4744]: E1003 16:26:49.892180 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:49 crc kubenswrapper[4744]: I1003 16:26:49.891965 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:49 crc kubenswrapper[4744]: I1003 16:26:49.891955 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:49 crc kubenswrapper[4744]: E1003 16:26:49.892406 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:49 crc kubenswrapper[4744]: E1003 16:26:49.892572 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:49 crc kubenswrapper[4744]: E1003 16:26:49.892741 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:50 crc kubenswrapper[4744]: E1003 16:26:50.007919 4744 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 03 16:26:51 crc kubenswrapper[4744]: I1003 16:26:51.891909 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:51 crc kubenswrapper[4744]: E1003 16:26:51.892048 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:51 crc kubenswrapper[4744]: I1003 16:26:51.892617 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:51 crc kubenswrapper[4744]: E1003 16:26:51.892698 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:51 crc kubenswrapper[4744]: I1003 16:26:51.892809 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:51 crc kubenswrapper[4744]: I1003 16:26:51.892862 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:51 crc kubenswrapper[4744]: E1003 16:26:51.892961 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:51 crc kubenswrapper[4744]: I1003 16:26:51.893604 4744 scope.go:117] "RemoveContainer" containerID="e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb" Oct 03 16:26:51 crc kubenswrapper[4744]: E1003 16:26:51.894023 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:52 crc kubenswrapper[4744]: I1003 16:26:52.618133 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovnkube-controller/3.log" Oct 03 16:26:52 crc kubenswrapper[4744]: I1003 16:26:52.621245 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerStarted","Data":"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0"} Oct 03 16:26:52 crc kubenswrapper[4744]: I1003 16:26:52.621799 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:26:52 crc kubenswrapper[4744]: I1003 16:26:52.815421 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podStartSLOduration=116.815394465 podStartE2EDuration="1m56.815394465s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:26:52.660107747 +0000 UTC m=+138.939983673" watchObservedRunningTime="2025-10-03 16:26:52.815394465 +0000 UTC m=+139.095270361" Oct 03 16:26:52 crc kubenswrapper[4744]: I1003 16:26:52.816361 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4f7dk"] Oct 03 16:26:52 crc kubenswrapper[4744]: I1003 16:26:52.816469 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:52 crc kubenswrapper[4744]: E1003 16:26:52.816621 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:53 crc kubenswrapper[4744]: I1003 16:26:53.891679 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:53 crc kubenswrapper[4744]: I1003 16:26:53.891823 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:53 crc kubenswrapper[4744]: E1003 16:26:53.891854 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:53 crc kubenswrapper[4744]: I1003 16:26:53.891700 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:53 crc kubenswrapper[4744]: E1003 16:26:53.892048 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:53 crc kubenswrapper[4744]: E1003 16:26:53.892058 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:54 crc kubenswrapper[4744]: I1003 16:26:54.891539 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:54 crc kubenswrapper[4744]: E1003 16:26:54.893423 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:55 crc kubenswrapper[4744]: E1003 16:26:55.008631 4744 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 03 16:26:55 crc kubenswrapper[4744]: I1003 16:26:55.891813 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:55 crc kubenswrapper[4744]: I1003 16:26:55.892258 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:55 crc kubenswrapper[4744]: I1003 16:26:55.891750 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:55 crc kubenswrapper[4744]: E1003 16:26:55.892455 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:55 crc kubenswrapper[4744]: E1003 16:26:55.892895 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:55 crc kubenswrapper[4744]: E1003 16:26:55.893273 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:56 crc kubenswrapper[4744]: I1003 16:26:56.891735 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:56 crc kubenswrapper[4744]: E1003 16:26:56.892852 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:57 crc kubenswrapper[4744]: I1003 16:26:57.891690 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:57 crc kubenswrapper[4744]: I1003 16:26:57.891751 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:57 crc kubenswrapper[4744]: E1003 16:26:57.891844 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:57 crc kubenswrapper[4744]: I1003 16:26:57.891954 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:57 crc kubenswrapper[4744]: E1003 16:26:57.892096 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:57 crc kubenswrapper[4744]: E1003 16:26:57.892336 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:26:58 crc kubenswrapper[4744]: I1003 16:26:58.891611 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:26:58 crc kubenswrapper[4744]: E1003 16:26:58.891764 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4f7dk" podUID="1008bf15-63d7-45f4-b42b-a4680b7dc232" Oct 03 16:26:59 crc kubenswrapper[4744]: I1003 16:26:59.890846 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:26:59 crc kubenswrapper[4744]: I1003 16:26:59.890854 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:26:59 crc kubenswrapper[4744]: E1003 16:26:59.890988 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 16:26:59 crc kubenswrapper[4744]: I1003 16:26:59.890866 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:26:59 crc kubenswrapper[4744]: E1003 16:26:59.891091 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 16:26:59 crc kubenswrapper[4744]: E1003 16:26:59.891116 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 16:27:00 crc kubenswrapper[4744]: I1003 16:27:00.891201 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:27:00 crc kubenswrapper[4744]: I1003 16:27:00.894331 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 03 16:27:00 crc kubenswrapper[4744]: I1003 16:27:00.895803 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 03 16:27:01 crc kubenswrapper[4744]: I1003 16:27:01.891510 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:27:01 crc kubenswrapper[4744]: I1003 16:27:01.891508 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:27:01 crc kubenswrapper[4744]: I1003 16:27:01.891542 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:27:01 crc kubenswrapper[4744]: I1003 16:27:01.894238 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 03 16:27:01 crc kubenswrapper[4744]: I1003 16:27:01.894253 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 03 16:27:01 crc kubenswrapper[4744]: I1003 16:27:01.894279 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 03 16:27:01 crc kubenswrapper[4744]: I1003 16:27:01.894310 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 03 16:27:02 crc kubenswrapper[4744]: I1003 16:27:02.750894 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:02 crc kubenswrapper[4744]: E1003 16:27:02.751063 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:29:04.75102473 +0000 UTC m=+271.030900626 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:02 crc kubenswrapper[4744]: I1003 16:27:02.751150 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:27:02 crc kubenswrapper[4744]: I1003 16:27:02.751206 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:27:02 crc kubenswrapper[4744]: I1003 16:27:02.752385 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:27:02 crc kubenswrapper[4744]: I1003 16:27:02.757613 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:27:02 crc kubenswrapper[4744]: I1003 16:27:02.806427 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 16:27:02 crc kubenswrapper[4744]: I1003 16:27:02.851954 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:27:02 crc kubenswrapper[4744]: I1003 16:27:02.851992 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:27:02 crc kubenswrapper[4744]: I1003 16:27:02.856016 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:27:02 crc kubenswrapper[4744]: I1003 16:27:02.856394 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:27:03 crc kubenswrapper[4744]: I1003 16:27:03.113560 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:27:03 crc kubenswrapper[4744]: I1003 16:27:03.120068 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 16:27:03 crc kubenswrapper[4744]: W1003 16:27:03.318778 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-abb72505dfbd11bdcdafcbbb139350e1e215f973abbc8670faf093cf6d6938ee WatchSource:0}: Error finding container abb72505dfbd11bdcdafcbbb139350e1e215f973abbc8670faf093cf6d6938ee: Status 404 returned error can't find the container with id abb72505dfbd11bdcdafcbbb139350e1e215f973abbc8670faf093cf6d6938ee Oct 03 16:27:03 crc kubenswrapper[4744]: W1003 16:27:03.359723 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-13736bff028a003516daf9d489d1bfbd5a244bd696cbed0328b660638b9f099d WatchSource:0}: Error finding container 13736bff028a003516daf9d489d1bfbd5a244bd696cbed0328b660638b9f099d: Status 404 returned error can't find the container with id 13736bff028a003516daf9d489d1bfbd5a244bd696cbed0328b660638b9f099d Oct 03 16:27:03 crc kubenswrapper[4744]: I1003 16:27:03.660598 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7c9c749e19f9a721aec8dc3d32144a31bec6c09abfe66f4824085276e4440e0b"} Oct 03 16:27:03 crc kubenswrapper[4744]: I1003 16:27:03.661238 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"13736bff028a003516daf9d489d1bfbd5a244bd696cbed0328b660638b9f099d"} Oct 03 16:27:03 crc kubenswrapper[4744]: I1003 16:27:03.664758 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7dc8ef79ed06e4d55a01d3fbc0891c16cbae11975978ce2a3fc35355ac39ab5a"} Oct 03 16:27:03 crc kubenswrapper[4744]: I1003 16:27:03.664808 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"abb72505dfbd11bdcdafcbbb139350e1e215f973abbc8670faf093cf6d6938ee"} Oct 03 16:27:03 crc kubenswrapper[4744]: I1003 16:27:03.665196 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:27:03 crc kubenswrapper[4744]: I1003 16:27:03.668062 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"19c6b3a886ced044145ddf3a410a3ca0d78ef406cce533e523c72c8b53ba5021"} Oct 03 16:27:03 crc kubenswrapper[4744]: I1003 16:27:03.668119 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4dc6645b031ad71c73486c69e8d5798b2719e4d102b3cf943124a017163eca18"} Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.669063 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.669171 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.745705 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.789645 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2pjgv"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.790746 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.791024 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-btrfh"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.791907 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" Oct 03 16:27:04 crc kubenswrapper[4744]: W1003 16:27:04.793577 4744 reflector.go:561] object-"openshift-apiserver"/"image-import-ca": failed to list *v1.ConfigMap: configmaps "image-import-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 03 16:27:04 crc kubenswrapper[4744]: E1003 16:27:04.793645 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"image-import-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"image-import-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 03 16:27:04 crc kubenswrapper[4744]: W1003 16:27:04.794257 4744 reflector.go:561] object-"openshift-apiserver"/"etcd-serving-ca": failed to list *v1.ConfigMap: configmaps "etcd-serving-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 03 16:27:04 crc kubenswrapper[4744]: E1003 16:27:04.794282 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"etcd-serving-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"etcd-serving-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.794354 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.795176 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.795210 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.795302 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.796418 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.796435 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.796685 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.797147 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.797170 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.797455 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.798086 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.798838 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.801644 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.802259 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.805684 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.806085 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.806165 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.806367 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.811518 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.811897 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.812087 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.812331 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.812490 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.812821 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.813051 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.813225 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.813536 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.813731 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.813890 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.814057 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.814268 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.814426 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.814651 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.815416 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xb6fx"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.815908 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.816036 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.816362 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.816752 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.818829 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.820856 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.823174 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.823321 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.825719 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.825771 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.825784 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.826218 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.826259 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.826314 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.826387 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.826462 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.826584 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.826635 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.826952 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.827238 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.827727 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.827856 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.827752 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.839636 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.841755 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-btrfh"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.846269 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2pjgv"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.846326 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xb6fx"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.881281 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.881392 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.881848 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.881897 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnnjb\" (UniqueName: \"kubernetes.io/projected/3f067415-c993-4ab6-9eda-84c315a8f4f9-kube-api-access-mnnjb\") pod \"openshift-apiserver-operator-796bbdcf4f-b8wvx\" (UID: \"3f067415-c993-4ab6-9eda-84c315a8f4f9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.881943 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b81e969d-bc65-4b6a-aaa3-ced704e3c744-serving-cert\") pod \"route-controller-manager-6576b87f9c-xj27k\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.881974 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-audit-dir\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.881998 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvqgt\" (UniqueName: \"kubernetes.io/projected/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-kube-api-access-kvqgt\") pod \"controller-manager-879f6c89f-xb6fx\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882022 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-node-pullsecrets\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882056 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-audit\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882198 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b81e969d-bc65-4b6a-aaa3-ced704e3c744-config\") pod \"route-controller-manager-6576b87f9c-xj27k\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882269 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c-images\") pod \"machine-api-operator-5694c8668f-btrfh\" (UID: \"23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882301 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-encryption-config\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882344 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4-auth-proxy-config\") pod \"machine-approver-56656f9798-t84ms\" (UID: \"e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882371 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-serving-cert\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882398 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-etcd-client\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882432 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf2cx\" (UniqueName: \"kubernetes.io/projected/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-kube-api-access-wf2cx\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882506 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f067415-c993-4ab6-9eda-84c315a8f4f9-config\") pod \"openshift-apiserver-operator-796bbdcf4f-b8wvx\" (UID: \"3f067415-c993-4ab6-9eda-84c315a8f4f9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882558 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882592 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-config\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882646 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-audit-policies\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882678 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882711 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b81e969d-bc65-4b6a-aaa3-ced704e3c744-client-ca\") pod \"route-controller-manager-6576b87f9c-xj27k\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882748 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4-config\") pod \"machine-approver-56656f9798-t84ms\" (UID: \"e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882776 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xb6fx\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882818 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm59n\" (UniqueName: \"kubernetes.io/projected/23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c-kube-api-access-xm59n\") pod \"machine-api-operator-5694c8668f-btrfh\" (UID: \"23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882855 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-btrfh\" (UID: \"23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882885 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7p99\" (UniqueName: \"kubernetes.io/projected/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-kube-api-access-z7p99\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882913 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-etcd-serving-ca\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.882959 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-serving-cert\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.883045 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-serving-cert\") pod \"controller-manager-879f6c89f-xb6fx\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.883085 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f067415-c993-4ab6-9eda-84c315a8f4f9-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-b8wvx\" (UID: \"3f067415-c993-4ab6-9eda-84c315a8f4f9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.883111 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-image-import-ca\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.883140 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-client-ca\") pod \"controller-manager-879f6c89f-xb6fx\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.883166 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-thbg6"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.883173 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kbcn\" (UniqueName: \"kubernetes.io/projected/b81e969d-bc65-4b6a-aaa3-ced704e3c744-kube-api-access-7kbcn\") pod \"route-controller-manager-6576b87f9c-xj27k\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.883679 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c-config\") pod \"machine-api-operator-5694c8668f-btrfh\" (UID: \"23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.883739 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jz5f\" (UniqueName: \"kubernetes.io/projected/e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4-kube-api-access-4jz5f\") pod \"machine-approver-56656f9798-t84ms\" (UID: \"e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.883776 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4-machine-approver-tls\") pod \"machine-approver-56656f9798-t84ms\" (UID: \"e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.883794 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-config\") pod \"controller-manager-879f6c89f-xb6fx\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.883824 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-etcd-client\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.883847 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-encryption-config\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.883863 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-thbg6" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.883865 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-audit-dir\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.884233 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-t9pf9"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.884734 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.885281 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6t5pf"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.885664 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.885774 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6t5pf" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.886417 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.886487 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.886595 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.886913 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.887599 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ksbgj"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.887996 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.888945 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.889138 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6jxqm"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.889895 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.890934 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.891114 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.891853 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.892217 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.892355 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.898121 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.898152 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.898198 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.898389 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.899231 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.899547 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.899846 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.899998 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.900100 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.900406 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.900479 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.900631 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.900716 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.900773 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.901030 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-v5snt"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.901457 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.901859 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.902813 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.902938 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.904569 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-946kz"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.905390 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.905696 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.906465 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.907110 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.907277 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.907335 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.907463 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.907539 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.907631 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.909174 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.909385 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.915439 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-6tzpn"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.916404 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.916651 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.917405 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.918059 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.918372 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.918465 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.918517 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.919361 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.920167 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.922216 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.922501 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.924712 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.924933 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.925545 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.925598 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.925711 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.925743 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.925816 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.926384 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.926539 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.926658 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.941736 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.946711 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.949315 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.966538 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.966580 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.967699 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94qmg"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.968300 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94qmg" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.968814 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.969192 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.969585 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.975199 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.976490 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.980827 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.981954 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.982735 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.982786 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.983247 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.984682 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnnjb\" (UniqueName: \"kubernetes.io/projected/3f067415-c993-4ab6-9eda-84c315a8f4f9-kube-api-access-mnnjb\") pod \"openshift-apiserver-operator-796bbdcf4f-b8wvx\" (UID: \"3f067415-c993-4ab6-9eda-84c315a8f4f9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.984764 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.984822 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqvrf\" (UniqueName: \"kubernetes.io/projected/4b772820-bbb4-4dec-937d-6502175fae6a-kube-api-access-rqvrf\") pod \"authentication-operator-69f744f599-t9pf9\" (UID: \"4b772820-bbb4-4dec-937d-6502175fae6a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.984851 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b81e969d-bc65-4b6a-aaa3-ced704e3c744-serving-cert\") pod \"route-controller-manager-6576b87f9c-xj27k\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.984871 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b550c60f-b9e0-44a5-938f-de24afa6c123-service-ca-bundle\") pod \"router-default-5444994796-946kz\" (UID: \"b550c60f-b9e0-44a5-938f-de24afa6c123\") " pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.984893 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dad9df42-c989-4fdf-8f19-ed7873979522-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wc4mg\" (UID: \"dad9df42-c989-4fdf-8f19-ed7873979522\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.984911 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-trusted-ca-bundle\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.984928 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b772820-bbb4-4dec-937d-6502175fae6a-service-ca-bundle\") pod \"authentication-operator-69f744f599-t9pf9\" (UID: \"4b772820-bbb4-4dec-937d-6502175fae6a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.984949 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-audit-dir\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.984970 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-console-config\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.984999 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvqgt\" (UniqueName: \"kubernetes.io/projected/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-kube-api-access-kvqgt\") pod \"controller-manager-879f6c89f-xb6fx\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985028 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-node-pullsecrets\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985052 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-audit\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985081 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985102 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b550c60f-b9e0-44a5-938f-de24afa6c123-default-certificate\") pod \"router-default-5444994796-946kz\" (UID: \"b550c60f-b9e0-44a5-938f-de24afa6c123\") " pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985123 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhsds\" (UniqueName: \"kubernetes.io/projected/f4865f7c-50e4-4159-8c57-42c69bd4efee-kube-api-access-jhsds\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985144 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4f39c88b-8d2a-4a19-921e-14d8d172ad20-audit-dir\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985167 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b81e969d-bc65-4b6a-aaa3-ced704e3c744-config\") pod \"route-controller-manager-6576b87f9c-xj27k\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985190 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f4865f7c-50e4-4159-8c57-42c69bd4efee-etcd-ca\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985211 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7fd8\" (UniqueName: \"kubernetes.io/projected/dad9df42-c989-4fdf-8f19-ed7873979522-kube-api-access-j7fd8\") pod \"machine-config-controller-84d6567774-wc4mg\" (UID: \"dad9df42-c989-4fdf-8f19-ed7873979522\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985229 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b772820-bbb4-4dec-937d-6502175fae6a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-t9pf9\" (UID: \"4b772820-bbb4-4dec-937d-6502175fae6a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985251 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-audit-policies\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985270 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985304 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n6qv\" (UniqueName: \"kubernetes.io/projected/4f39c88b-8d2a-4a19-921e-14d8d172ad20-kube-api-access-2n6qv\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985322 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b772820-bbb4-4dec-937d-6502175fae6a-config\") pod \"authentication-operator-69f744f599-t9pf9\" (UID: \"4b772820-bbb4-4dec-937d-6502175fae6a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985342 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c-images\") pod \"machine-api-operator-5694c8668f-btrfh\" (UID: \"23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985361 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-encryption-config\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985380 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d61086b0-b78f-4239-8d9a-49aba28c5820-console-serving-cert\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985400 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jqnn\" (UniqueName: \"kubernetes.io/projected/d61086b0-b78f-4239-8d9a-49aba28c5820-kube-api-access-6jqnn\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985425 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b550c60f-b9e0-44a5-938f-de24afa6c123-metrics-certs\") pod \"router-default-5444994796-946kz\" (UID: \"b550c60f-b9e0-44a5-938f-de24afa6c123\") " pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985442 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdczj\" (UniqueName: \"kubernetes.io/projected/b550c60f-b9e0-44a5-938f-de24afa6c123-kube-api-access-hdczj\") pod \"router-default-5444994796-946kz\" (UID: \"b550c60f-b9e0-44a5-938f-de24afa6c123\") " pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985465 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4-auth-proxy-config\") pod \"machine-approver-56656f9798-t84ms\" (UID: \"e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.985537 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-serving-cert\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.986155 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.987673 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-audit-dir\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.988325 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c-images\") pod \"machine-api-operator-5694c8668f-btrfh\" (UID: \"23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.988525 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-node-pullsecrets\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.988870 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-audit\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.989118 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4-auth-proxy-config\") pod \"machine-approver-56656f9798-t84ms\" (UID: \"e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.991599 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b81e969d-bc65-4b6a-aaa3-ced704e3c744-config\") pod \"route-controller-manager-6576b87f9c-xj27k\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.993051 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dfzqw"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.993984 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-dfzqw" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.994917 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-etcd-client\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.994956 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f4865f7c-50e4-4159-8c57-42c69bd4efee-etcd-client\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.994994 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf2cx\" (UniqueName: \"kubernetes.io/projected/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-kube-api-access-wf2cx\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995016 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b550c60f-b9e0-44a5-938f-de24afa6c123-stats-auth\") pod \"router-default-5444994796-946kz\" (UID: \"b550c60f-b9e0-44a5-938f-de24afa6c123\") " pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995042 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b772820-bbb4-4dec-937d-6502175fae6a-serving-cert\") pod \"authentication-operator-69f744f599-t9pf9\" (UID: \"4b772820-bbb4-4dec-937d-6502175fae6a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995071 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f067415-c993-4ab6-9eda-84c315a8f4f9-config\") pod \"openshift-apiserver-operator-796bbdcf4f-b8wvx\" (UID: \"3f067415-c993-4ab6-9eda-84c315a8f4f9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995091 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995113 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0b52a53-304a-4664-b8d7-9434f969dac3-config\") pod \"kube-apiserver-operator-766d6c64bb-w8sw7\" (UID: \"a0b52a53-304a-4664-b8d7-9434f969dac3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995135 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995157 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxq7v\" (UniqueName: \"kubernetes.io/projected/6ac85db3-283c-43f9-9925-b8a0f8354181-kube-api-access-hxq7v\") pod \"downloads-7954f5f757-thbg6\" (UID: \"6ac85db3-283c-43f9-9925-b8a0f8354181\") " pod="openshift-console/downloads-7954f5f757-thbg6" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995179 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f62k2\" (UniqueName: \"kubernetes.io/projected/37928f66-a94f-4241-b984-e0729ee6dbf0-kube-api-access-f62k2\") pod \"openshift-config-operator-7777fb866f-jf2dh\" (UID: \"37928f66-a94f-4241-b984-e0729ee6dbf0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995203 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-config\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995225 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-audit-policies\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995246 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995265 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0c1475e-1086-40fc-800e-91f5156c92f9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2kpgr\" (UID: \"a0c1475e-1086-40fc-800e-91f5156c92f9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995286 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpz9s\" (UniqueName: \"kubernetes.io/projected/a0c1475e-1086-40fc-800e-91f5156c92f9-kube-api-access-hpz9s\") pod \"cluster-image-registry-operator-dc59b4c8b-2kpgr\" (UID: \"a0c1475e-1086-40fc-800e-91f5156c92f9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995309 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b81e969d-bc65-4b6a-aaa3-ced704e3c744-client-ca\") pod \"route-controller-manager-6576b87f9c-xj27k\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995328 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f4865f7c-50e4-4159-8c57-42c69bd4efee-etcd-service-ca\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995349 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995379 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4-config\") pod \"machine-approver-56656f9798-t84ms\" (UID: \"e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995416 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xb6fx\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995446 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm59n\" (UniqueName: \"kubernetes.io/projected/23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c-kube-api-access-xm59n\") pod \"machine-api-operator-5694c8668f-btrfh\" (UID: \"23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995473 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc4br\" (UniqueName: \"kubernetes.io/projected/b2a7d1a9-0abe-4150-8c23-c9f592467ba5-kube-api-access-wc4br\") pod \"migrator-59844c95c7-6t5pf\" (UID: \"b2a7d1a9-0abe-4150-8c23-c9f592467ba5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6t5pf" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995522 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-btrfh\" (UID: \"23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995554 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7p99\" (UniqueName: \"kubernetes.io/projected/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-kube-api-access-z7p99\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995577 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-service-ca\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995602 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-etcd-serving-ca\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995619 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-serving-cert\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995640 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dad9df42-c989-4fdf-8f19-ed7873979522-proxy-tls\") pod \"machine-config-controller-84d6567774-wc4mg\" (UID: \"dad9df42-c989-4fdf-8f19-ed7873979522\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995660 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-serving-cert\") pod \"controller-manager-879f6c89f-xb6fx\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995676 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0b52a53-304a-4664-b8d7-9434f969dac3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-w8sw7\" (UID: \"a0b52a53-304a-4664-b8d7-9434f969dac3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995697 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995716 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/37928f66-a94f-4241-b984-e0729ee6dbf0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-jf2dh\" (UID: \"37928f66-a94f-4241-b984-e0729ee6dbf0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995740 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f067415-c993-4ab6-9eda-84c315a8f4f9-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-b8wvx\" (UID: \"3f067415-c993-4ab6-9eda-84c315a8f4f9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995756 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-image-import-ca\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995773 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a0c1475e-1086-40fc-800e-91f5156c92f9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2kpgr\" (UID: \"a0c1475e-1086-40fc-800e-91f5156c92f9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995794 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-client-ca\") pod \"controller-manager-879f6c89f-xb6fx\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995812 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-oauth-serving-cert\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995829 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37928f66-a94f-4241-b984-e0729ee6dbf0-serving-cert\") pod \"openshift-config-operator-7777fb866f-jf2dh\" (UID: \"37928f66-a94f-4241-b984-e0729ee6dbf0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995859 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kbcn\" (UniqueName: \"kubernetes.io/projected/b81e969d-bc65-4b6a-aaa3-ced704e3c744-kube-api-access-7kbcn\") pod \"route-controller-manager-6576b87f9c-xj27k\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995878 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c-config\") pod \"machine-api-operator-5694c8668f-btrfh\" (UID: \"23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995895 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0b52a53-304a-4664-b8d7-9434f969dac3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-w8sw7\" (UID: \"a0b52a53-304a-4664-b8d7-9434f969dac3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995921 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jz5f\" (UniqueName: \"kubernetes.io/projected/e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4-kube-api-access-4jz5f\") pod \"machine-approver-56656f9798-t84ms\" (UID: \"e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995937 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4865f7c-50e4-4159-8c57-42c69bd4efee-config\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995959 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.995983 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4-machine-approver-tls\") pod \"machine-approver-56656f9798-t84ms\" (UID: \"e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.996000 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-config\") pod \"controller-manager-879f6c89f-xb6fx\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.996024 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a0c1475e-1086-40fc-800e-91f5156c92f9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2kpgr\" (UID: \"a0c1475e-1086-40fc-800e-91f5156c92f9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.996042 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.996086 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-etcd-client\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.996104 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.996124 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-encryption-config\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.996142 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-audit-dir\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.996160 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.996181 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.996198 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4865f7c-50e4-4159-8c57-42c69bd4efee-serving-cert\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.996215 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d61086b0-b78f-4239-8d9a-49aba28c5820-console-oauth-config\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.996233 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.996963 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.997048 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f067415-c993-4ab6-9eda-84c315a8f4f9-config\") pod \"openshift-apiserver-operator-796bbdcf4f-b8wvx\" (UID: \"3f067415-c993-4ab6-9eda-84c315a8f4f9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.997109 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-k2tzp"] Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.997698 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-audit-policies\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.998227 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.998552 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-client-ca\") pod \"controller-manager-879f6c89f-xb6fx\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.999046 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b81e969d-bc65-4b6a-aaa3-ced704e3c744-client-ca\") pod \"route-controller-manager-6576b87f9c-xj27k\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:04 crc kubenswrapper[4744]: I1003 16:27:04.999336 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-config\") pod \"controller-manager-879f6c89f-xb6fx\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.001708 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.001754 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.001803 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-audit-dir\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.003143 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-config\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.004222 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-etcd-client\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.004822 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.005413 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xb6fx\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.005549 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.005898 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c-config\") pod \"machine-api-operator-5694c8668f-btrfh\" (UID: \"23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.006619 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmzm9"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.006709 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-serving-cert\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.006900 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.007937 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmzm9" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.008446 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-nbsh6"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.009407 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-nbsh6" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.010586 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-serving-cert\") pod \"controller-manager-879f6c89f-xb6fx\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.011079 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.012615 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f067415-c993-4ab6-9eda-84c315a8f4f9-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-b8wvx\" (UID: \"3f067415-c993-4ab6-9eda-84c315a8f4f9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.013341 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4-machine-approver-tls\") pod \"machine-approver-56656f9798-t84ms\" (UID: \"e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.014900 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-encryption-config\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.016443 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b81e969d-bc65-4b6a-aaa3-ced704e3c744-serving-cert\") pod \"route-controller-manager-6576b87f9c-xj27k\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.017724 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4-config\") pod \"machine-approver-56656f9798-t84ms\" (UID: \"e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.020380 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.021334 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-btrfh\" (UID: \"23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.024677 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.025460 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dpx8q"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.025968 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dpx8q" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.025962 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.026040 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.027922 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.033364 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-etcd-client\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.039913 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.040086 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-encryption-config\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.042997 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-serving-cert\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.054154 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.054592 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.055468 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.057217 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9hm6h"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.065148 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.065747 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.076392 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.078304 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.078458 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.078921 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.078326 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-d6zzl"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.083831 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.084211 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.085374 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.087312 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-qmddz"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.087743 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.087819 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.088099 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-s96bg"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.088345 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-qmddz" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.089249 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-v5snt"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.089274 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.089286 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.089298 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6t5pf"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.089387 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-s96bg" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.089890 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-k2tzp"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.091387 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-t9pf9"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.093837 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.094818 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.096690 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6jxqm"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.096902 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-service-ca\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.096949 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dad9df42-c989-4fdf-8f19-ed7873979522-proxy-tls\") pod \"machine-config-controller-84d6567774-wc4mg\" (UID: \"dad9df42-c989-4fdf-8f19-ed7873979522\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.096973 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0b52a53-304a-4664-b8d7-9434f969dac3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-w8sw7\" (UID: \"a0b52a53-304a-4664-b8d7-9434f969dac3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097001 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a0c1475e-1086-40fc-800e-91f5156c92f9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2kpgr\" (UID: \"a0c1475e-1086-40fc-800e-91f5156c92f9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097021 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097042 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/37928f66-a94f-4241-b984-e0729ee6dbf0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-jf2dh\" (UID: \"37928f66-a94f-4241-b984-e0729ee6dbf0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097063 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-oauth-serving-cert\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097081 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37928f66-a94f-4241-b984-e0729ee6dbf0-serving-cert\") pod \"openshift-config-operator-7777fb866f-jf2dh\" (UID: \"37928f66-a94f-4241-b984-e0729ee6dbf0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097110 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0b52a53-304a-4664-b8d7-9434f969dac3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-w8sw7\" (UID: \"a0b52a53-304a-4664-b8d7-9434f969dac3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097137 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4865f7c-50e4-4159-8c57-42c69bd4efee-config\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097158 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097179 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a0c1475e-1086-40fc-800e-91f5156c92f9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2kpgr\" (UID: \"a0c1475e-1086-40fc-800e-91f5156c92f9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097198 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097221 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097244 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097265 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4865f7c-50e4-4159-8c57-42c69bd4efee-serving-cert\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097287 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d61086b0-b78f-4239-8d9a-49aba28c5820-console-oauth-config\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097310 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097344 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097363 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqvrf\" (UniqueName: \"kubernetes.io/projected/4b772820-bbb4-4dec-937d-6502175fae6a-kube-api-access-rqvrf\") pod \"authentication-operator-69f744f599-t9pf9\" (UID: \"4b772820-bbb4-4dec-937d-6502175fae6a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097384 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b550c60f-b9e0-44a5-938f-de24afa6c123-service-ca-bundle\") pod \"router-default-5444994796-946kz\" (UID: \"b550c60f-b9e0-44a5-938f-de24afa6c123\") " pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097406 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dad9df42-c989-4fdf-8f19-ed7873979522-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wc4mg\" (UID: \"dad9df42-c989-4fdf-8f19-ed7873979522\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097426 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-console-config\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097442 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-trusted-ca-bundle\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097461 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b772820-bbb4-4dec-937d-6502175fae6a-service-ca-bundle\") pod \"authentication-operator-69f744f599-t9pf9\" (UID: \"4b772820-bbb4-4dec-937d-6502175fae6a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097505 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097523 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b550c60f-b9e0-44a5-938f-de24afa6c123-default-certificate\") pod \"router-default-5444994796-946kz\" (UID: \"b550c60f-b9e0-44a5-938f-de24afa6c123\") " pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097543 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f4865f7c-50e4-4159-8c57-42c69bd4efee-etcd-ca\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097559 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhsds\" (UniqueName: \"kubernetes.io/projected/f4865f7c-50e4-4159-8c57-42c69bd4efee-kube-api-access-jhsds\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097576 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4f39c88b-8d2a-4a19-921e-14d8d172ad20-audit-dir\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097597 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-audit-policies\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097613 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097637 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n6qv\" (UniqueName: \"kubernetes.io/projected/4f39c88b-8d2a-4a19-921e-14d8d172ad20-kube-api-access-2n6qv\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097655 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7fd8\" (UniqueName: \"kubernetes.io/projected/dad9df42-c989-4fdf-8f19-ed7873979522-kube-api-access-j7fd8\") pod \"machine-config-controller-84d6567774-wc4mg\" (UID: \"dad9df42-c989-4fdf-8f19-ed7873979522\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097673 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b772820-bbb4-4dec-937d-6502175fae6a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-t9pf9\" (UID: \"4b772820-bbb4-4dec-937d-6502175fae6a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097692 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d61086b0-b78f-4239-8d9a-49aba28c5820-console-serving-cert\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097708 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jqnn\" (UniqueName: \"kubernetes.io/projected/d61086b0-b78f-4239-8d9a-49aba28c5820-kube-api-access-6jqnn\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097726 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b772820-bbb4-4dec-937d-6502175fae6a-config\") pod \"authentication-operator-69f744f599-t9pf9\" (UID: \"4b772820-bbb4-4dec-937d-6502175fae6a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097744 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f4865f7c-50e4-4159-8c57-42c69bd4efee-etcd-client\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097762 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b550c60f-b9e0-44a5-938f-de24afa6c123-metrics-certs\") pod \"router-default-5444994796-946kz\" (UID: \"b550c60f-b9e0-44a5-938f-de24afa6c123\") " pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097779 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdczj\" (UniqueName: \"kubernetes.io/projected/b550c60f-b9e0-44a5-938f-de24afa6c123-kube-api-access-hdczj\") pod \"router-default-5444994796-946kz\" (UID: \"b550c60f-b9e0-44a5-938f-de24afa6c123\") " pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097802 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b550c60f-b9e0-44a5-938f-de24afa6c123-stats-auth\") pod \"router-default-5444994796-946kz\" (UID: \"b550c60f-b9e0-44a5-938f-de24afa6c123\") " pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097823 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0b52a53-304a-4664-b8d7-9434f969dac3-config\") pod \"kube-apiserver-operator-766d6c64bb-w8sw7\" (UID: \"a0b52a53-304a-4664-b8d7-9434f969dac3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097841 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097862 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b772820-bbb4-4dec-937d-6502175fae6a-serving-cert\") pod \"authentication-operator-69f744f599-t9pf9\" (UID: \"4b772820-bbb4-4dec-937d-6502175fae6a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097880 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0c1475e-1086-40fc-800e-91f5156c92f9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2kpgr\" (UID: \"a0c1475e-1086-40fc-800e-91f5156c92f9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097899 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpz9s\" (UniqueName: \"kubernetes.io/projected/a0c1475e-1086-40fc-800e-91f5156c92f9-kube-api-access-hpz9s\") pod \"cluster-image-registry-operator-dc59b4c8b-2kpgr\" (UID: \"a0c1475e-1086-40fc-800e-91f5156c92f9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097916 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxq7v\" (UniqueName: \"kubernetes.io/projected/6ac85db3-283c-43f9-9925-b8a0f8354181-kube-api-access-hxq7v\") pod \"downloads-7954f5f757-thbg6\" (UID: \"6ac85db3-283c-43f9-9925-b8a0f8354181\") " pod="openshift-console/downloads-7954f5f757-thbg6" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097939 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f62k2\" (UniqueName: \"kubernetes.io/projected/37928f66-a94f-4241-b984-e0729ee6dbf0-kube-api-access-f62k2\") pod \"openshift-config-operator-7777fb866f-jf2dh\" (UID: \"37928f66-a94f-4241-b984-e0729ee6dbf0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097958 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f4865f7c-50e4-4159-8c57-42c69bd4efee-etcd-service-ca\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.097976 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.098002 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc4br\" (UniqueName: \"kubernetes.io/projected/b2a7d1a9-0abe-4150-8c23-c9f592467ba5-kube-api-access-wc4br\") pod \"migrator-59844c95c7-6t5pf\" (UID: \"b2a7d1a9-0abe-4150-8c23-c9f592467ba5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6t5pf" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.098046 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/37928f66-a94f-4241-b984-e0729ee6dbf0-available-featuregates\") pod \"openshift-config-operator-7777fb866f-jf2dh\" (UID: \"37928f66-a94f-4241-b984-e0729ee6dbf0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.098164 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.098273 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-service-ca\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.099087 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4865f7c-50e4-4159-8c57-42c69bd4efee-config\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.099129 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-oauth-serving-cert\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.099152 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.100369 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f4865f7c-50e4-4159-8c57-42c69bd4efee-etcd-ca\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.100404 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmzm9"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.099817 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-audit-policies\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.101605 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-nbsh6"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.101715 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b772820-bbb4-4dec-937d-6502175fae6a-service-ca-bundle\") pod \"authentication-operator-69f744f599-t9pf9\" (UID: \"4b772820-bbb4-4dec-937d-6502175fae6a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.102085 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dad9df42-c989-4fdf-8f19-ed7873979522-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wc4mg\" (UID: \"dad9df42-c989-4fdf-8f19-ed7873979522\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.102715 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-console-config\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.103298 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b550c60f-b9e0-44a5-938f-de24afa6c123-service-ca-bundle\") pod \"router-default-5444994796-946kz\" (UID: \"b550c60f-b9e0-44a5-938f-de24afa6c123\") " pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.103466 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37928f66-a94f-4241-b984-e0729ee6dbf0-serving-cert\") pod \"openshift-config-operator-7777fb866f-jf2dh\" (UID: \"37928f66-a94f-4241-b984-e0729ee6dbf0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.103960 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/b550c60f-b9e0-44a5-938f-de24afa6c123-default-certificate\") pod \"router-default-5444994796-946kz\" (UID: \"b550c60f-b9e0-44a5-938f-de24afa6c123\") " pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.104436 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b772820-bbb4-4dec-937d-6502175fae6a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-t9pf9\" (UID: \"4b772820-bbb4-4dec-937d-6502175fae6a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.104456 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0b52a53-304a-4664-b8d7-9434f969dac3-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-w8sw7\" (UID: \"a0b52a53-304a-4664-b8d7-9434f969dac3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.104663 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-trusted-ca-bundle\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.104691 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.104725 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.105508 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.105865 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f4865f7c-50e4-4159-8c57-42c69bd4efee-etcd-service-ca\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.105979 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.106745 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.106800 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4f39c88b-8d2a-4a19-921e-14d8d172ad20-audit-dir\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.106993 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d61086b0-b78f-4239-8d9a-49aba28c5820-console-oauth-config\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.107077 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.107634 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dfzqw"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.107945 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-thbg6"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.108235 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b772820-bbb4-4dec-937d-6502175fae6a-serving-cert\") pod \"authentication-operator-69f744f599-t9pf9\" (UID: \"4b772820-bbb4-4dec-937d-6502175fae6a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.109447 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b772820-bbb4-4dec-937d-6502175fae6a-config\") pod \"authentication-operator-69f744f599-t9pf9\" (UID: \"4b772820-bbb4-4dec-937d-6502175fae6a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.109662 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.110015 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a0c1475e-1086-40fc-800e-91f5156c92f9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2kpgr\" (UID: \"a0c1475e-1086-40fc-800e-91f5156c92f9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.111194 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d61086b0-b78f-4239-8d9a-49aba28c5820-console-serving-cert\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.111521 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b550c60f-b9e0-44a5-938f-de24afa6c123-metrics-certs\") pod \"router-default-5444994796-946kz\" (UID: \"b550c60f-b9e0-44a5-938f-de24afa6c123\") " pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.111727 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.113027 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.113077 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.113510 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4865f7c-50e4-4159-8c57-42c69bd4efee-serving-cert\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.113609 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/b550c60f-b9e0-44a5-938f-de24afa6c123-stats-auth\") pod \"router-default-5444994796-946kz\" (UID: \"b550c60f-b9e0-44a5-938f-de24afa6c123\") " pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.114062 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0c1475e-1086-40fc-800e-91f5156c92f9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2kpgr\" (UID: \"a0c1475e-1086-40fc-800e-91f5156c92f9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.114166 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ksbgj"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.114195 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.114791 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0b52a53-304a-4664-b8d7-9434f969dac3-config\") pod \"kube-apiserver-operator-766d6c64bb-w8sw7\" (UID: \"a0b52a53-304a-4664-b8d7-9434f969dac3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.115234 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f4865f7c-50e4-4159-8c57-42c69bd4efee-etcd-client\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.115761 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.116483 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.116868 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-6tzpn"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.117810 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.118901 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.119893 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.121965 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.124149 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.125346 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.126459 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.127533 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dpx8q"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.128606 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9hm6h"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.130137 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-d6zzl"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.131402 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.132460 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.132996 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.133473 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94qmg"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.134564 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-qmddz"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.135613 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-s96bg"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.136556 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-dmld4"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.137342 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-dmld4" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.142131 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dad9df42-c989-4fdf-8f19-ed7873979522-proxy-tls\") pod \"machine-config-controller-84d6567774-wc4mg\" (UID: \"dad9df42-c989-4fdf-8f19-ed7873979522\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.152673 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.173512 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.193952 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.212927 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.234350 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.253682 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.274112 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.293149 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.314316 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.334154 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.353885 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.374195 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.434206 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.454047 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.473565 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.502037 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.514133 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.533588 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.554757 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.573789 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.594165 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.641179 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnnjb\" (UniqueName: \"kubernetes.io/projected/3f067415-c993-4ab6-9eda-84c315a8f4f9-kube-api-access-mnnjb\") pod \"openshift-apiserver-operator-796bbdcf4f-b8wvx\" (UID: \"3f067415-c993-4ab6-9eda-84c315a8f4f9\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.654290 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.662141 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvqgt\" (UniqueName: \"kubernetes.io/projected/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-kube-api-access-kvqgt\") pod \"controller-manager-879f6c89f-xb6fx\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.674083 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.674638 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.693746 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.713620 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.753457 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.766873 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf2cx\" (UniqueName: \"kubernetes.io/projected/90345bca-bb79-4ea3-9cc8-f6f8bebb3180-kube-api-access-wf2cx\") pod \"apiserver-7bbb656c7d-6vrs7\" (UID: \"90345bca-bb79-4ea3-9cc8-f6f8bebb3180\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.774580 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.794596 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.800714 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.814665 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.844085 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.854835 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.883486 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.896279 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jz5f\" (UniqueName: \"kubernetes.io/projected/e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4-kube-api-access-4jz5f\") pod \"machine-approver-56656f9798-t84ms\" (UID: \"e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.911839 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xb6fx"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.914044 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kbcn\" (UniqueName: \"kubernetes.io/projected/b81e969d-bc65-4b6a-aaa3-ced704e3c744-kube-api-access-7kbcn\") pod \"route-controller-manager-6576b87f9c-xj27k\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:05 crc kubenswrapper[4744]: W1003 16:27:05.922111 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bd33bcb_22df_4c14_9ccd_8661ff4615d5.slice/crio-32337321d3e006f6082f53d1339d25aa997bffc54d8d206c784b2c8c293b8199 WatchSource:0}: Error finding container 32337321d3e006f6082f53d1339d25aa997bffc54d8d206c784b2c8c293b8199: Status 404 returned error can't find the container with id 32337321d3e006f6082f53d1339d25aa997bffc54d8d206c784b2c8c293b8199 Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.929297 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm59n\" (UniqueName: \"kubernetes.io/projected/23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c-kube-api-access-xm59n\") pod \"machine-api-operator-5694c8668f-btrfh\" (UID: \"23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.952851 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.959732 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.963303 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7p99\" (UniqueName: \"kubernetes.io/projected/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-kube-api-access-z7p99\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.973909 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.982063 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7"] Oct 03 16:27:05 crc kubenswrapper[4744]: I1003 16:27:05.994151 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 03 16:27:06 crc kubenswrapper[4744]: E1003 16:27:05.998209 4744 configmap.go:193] Couldn't get configMap openshift-apiserver/image-import-ca: failed to sync configmap cache: timed out waiting for the condition Oct 03 16:27:06 crc kubenswrapper[4744]: E1003 16:27:05.998302 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-image-import-ca podName:5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7 nodeName:}" failed. No retries permitted until 2025-10-03 16:27:06.498280247 +0000 UTC m=+152.778156143 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "image-import-ca" (UniqueName: "kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-image-import-ca") pod "apiserver-76f77b778f-2pjgv" (UID: "5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7") : failed to sync configmap cache: timed out waiting for the condition Oct 03 16:27:06 crc kubenswrapper[4744]: E1003 16:27:06.002641 4744 configmap.go:193] Couldn't get configMap openshift-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Oct 03 16:27:06 crc kubenswrapper[4744]: E1003 16:27:06.002700 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-etcd-serving-ca podName:5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7 nodeName:}" failed. No retries permitted until 2025-10-03 16:27:06.502685497 +0000 UTC m=+152.782561393 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-etcd-serving-ca") pod "apiserver-76f77b778f-2pjgv" (UID: "5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7") : failed to sync configmap cache: timed out waiting for the condition Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.011165 4744 request.go:700] Waited for 1.003850369s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager-operator/configmaps?fieldSelector=metadata.name%3Dopenshift-controller-manager-operator-config&limit=500&resourceVersion=0 Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.017811 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.035561 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.054056 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.074274 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.091468 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.093956 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.114144 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.121464 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.140446 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 03 16:27:06 crc kubenswrapper[4744]: W1003 16:27:06.143258 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode50b3ffc_b380_4fd2_92f6_35b6b1dc5bd4.slice/crio-54f711ec437f911514ba609e60a4084b738ac39554d36a1876c2e52a120ea16b WatchSource:0}: Error finding container 54f711ec437f911514ba609e60a4084b738ac39554d36a1876c2e52a120ea16b: Status 404 returned error can't find the container with id 54f711ec437f911514ba609e60a4084b738ac39554d36a1876c2e52a120ea16b Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.152643 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.166908 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx"] Oct 03 16:27:06 crc kubenswrapper[4744]: W1003 16:27:06.174138 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f067415_c993_4ab6_9eda_84c315a8f4f9.slice/crio-5c615918d3f56d4f99ac8cf36c30dabe378dbee3f0696714e2ff8d3410145481 WatchSource:0}: Error finding container 5c615918d3f56d4f99ac8cf36c30dabe378dbee3f0696714e2ff8d3410145481: Status 404 returned error can't find the container with id 5c615918d3f56d4f99ac8cf36c30dabe378dbee3f0696714e2ff8d3410145481 Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.174421 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.194836 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.213635 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.216508 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k"] Oct 03 16:27:06 crc kubenswrapper[4744]: W1003 16:27:06.226529 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb81e969d_bc65_4b6a_aaa3_ced704e3c744.slice/crio-6c6b7c72dc428d1cc626f395f7191a3259f657fe9b6f94e03f1e858ce7638baa WatchSource:0}: Error finding container 6c6b7c72dc428d1cc626f395f7191a3259f657fe9b6f94e03f1e858ce7638baa: Status 404 returned error can't find the container with id 6c6b7c72dc428d1cc626f395f7191a3259f657fe9b6f94e03f1e858ce7638baa Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.234687 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.254198 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.279315 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.295024 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.303196 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-btrfh"] Oct 03 16:27:06 crc kubenswrapper[4744]: W1003 16:27:06.311213 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23b0d312_c1eb_47f2_a1ae_7fbcbc5ebb8c.slice/crio-890642a82c77f60a4d36a6a329c94d9ddcf954f5fc2df1948adc9da5a156f486 WatchSource:0}: Error finding container 890642a82c77f60a4d36a6a329c94d9ddcf954f5fc2df1948adc9da5a156f486: Status 404 returned error can't find the container with id 890642a82c77f60a4d36a6a329c94d9ddcf954f5fc2df1948adc9da5a156f486 Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.313958 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.333344 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.352635 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.374277 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.394202 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.414043 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.433524 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.454204 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.475140 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.493907 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.522689 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-etcd-serving-ca\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.523316 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-image-import-ca\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.525707 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.534059 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.553530 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.573190 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.593427 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.613811 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.634201 4744 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.653891 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.674242 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.689944 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" event={"ID":"e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4","Type":"ContainerStarted","Data":"7c4b94be026d03a35d4c6f21c0548fa8b6bd1cd3cd8e25ff28e454762c279bed"} Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.690009 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" event={"ID":"e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4","Type":"ContainerStarted","Data":"54f711ec437f911514ba609e60a4084b738ac39554d36a1876c2e52a120ea16b"} Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.692236 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" event={"ID":"b81e969d-bc65-4b6a-aaa3-ced704e3c744","Type":"ContainerStarted","Data":"de3f5609b11d80898bda549e00f708e4abf6cfcce1093c61b92861639829103e"} Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.692304 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" event={"ID":"b81e969d-bc65-4b6a-aaa3-ced704e3c744","Type":"ContainerStarted","Data":"6c6b7c72dc428d1cc626f395f7191a3259f657fe9b6f94e03f1e858ce7638baa"} Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.692944 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.694297 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.694905 4744 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-xj27k container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.694981 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" podUID="b81e969d-bc65-4b6a-aaa3-ced704e3c744" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.695433 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx" event={"ID":"3f067415-c993-4ab6-9eda-84c315a8f4f9","Type":"ContainerStarted","Data":"c40b75b29218e3aa23f8589d90e6ceb797dda5d65b6953b12d5b41d99807e7b6"} Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.695475 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx" event={"ID":"3f067415-c993-4ab6-9eda-84c315a8f4f9","Type":"ContainerStarted","Data":"5c615918d3f56d4f99ac8cf36c30dabe378dbee3f0696714e2ff8d3410145481"} Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.697114 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" event={"ID":"1bd33bcb-22df-4c14-9ccd-8661ff4615d5","Type":"ContainerStarted","Data":"4f37b00d6c7dc0c25e578e4f29ba45d009b18d77ffd56d78cae9f7dd5607a1bb"} Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.697140 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" event={"ID":"1bd33bcb-22df-4c14-9ccd-8661ff4615d5","Type":"ContainerStarted","Data":"32337321d3e006f6082f53d1339d25aa997bffc54d8d206c784b2c8c293b8199"} Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.697793 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.698807 4744 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-xb6fx container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.699644 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" podUID="1bd33bcb-22df-4c14-9ccd-8661ff4615d5" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.700187 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" event={"ID":"23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c","Type":"ContainerStarted","Data":"ed85c0cd7303938046752b7f7882847c40fac48c82bd232c16f27663520cc753"} Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.700252 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" event={"ID":"23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c","Type":"ContainerStarted","Data":"e599c42900b114b5ebc02f8d50f9a1cccbd2a6628eba906bc12834787ab607f2"} Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.700290 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" event={"ID":"23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c","Type":"ContainerStarted","Data":"890642a82c77f60a4d36a6a329c94d9ddcf954f5fc2df1948adc9da5a156f486"} Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.703295 4744 generic.go:334] "Generic (PLEG): container finished" podID="90345bca-bb79-4ea3-9cc8-f6f8bebb3180" containerID="a63586a920b2ef62df263e5ebb5b16f12a68ed88febfe7f4edc5dea80e2dd0b7" exitCode=0 Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.703335 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" event={"ID":"90345bca-bb79-4ea3-9cc8-f6f8bebb3180","Type":"ContainerDied","Data":"a63586a920b2ef62df263e5ebb5b16f12a68ed88febfe7f4edc5dea80e2dd0b7"} Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.703353 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" event={"ID":"90345bca-bb79-4ea3-9cc8-f6f8bebb3180","Type":"ContainerStarted","Data":"928273ff5808119b192c5e5e25a8cadaffd6817c2216de76e64815ade9401a58"} Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.713130 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.733666 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.753057 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.773967 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.793550 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.813564 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.834821 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.853523 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.874298 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.893324 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.914233 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.954876 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc4br\" (UniqueName: \"kubernetes.io/projected/b2a7d1a9-0abe-4150-8c23-c9f592467ba5-kube-api-access-wc4br\") pod \"migrator-59844c95c7-6t5pf\" (UID: \"b2a7d1a9-0abe-4150-8c23-c9f592467ba5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6t5pf" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.970799 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0b52a53-304a-4664-b8d7-9434f969dac3-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-w8sw7\" (UID: \"a0b52a53-304a-4664-b8d7-9434f969dac3\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7" Oct 03 16:27:06 crc kubenswrapper[4744]: I1003 16:27:06.993943 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqvrf\" (UniqueName: \"kubernetes.io/projected/4b772820-bbb4-4dec-937d-6502175fae6a-kube-api-access-rqvrf\") pod \"authentication-operator-69f744f599-t9pf9\" (UID: \"4b772820-bbb4-4dec-937d-6502175fae6a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.010375 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n6qv\" (UniqueName: \"kubernetes.io/projected/4f39c88b-8d2a-4a19-921e-14d8d172ad20-kube-api-access-2n6qv\") pod \"oauth-openshift-558db77b4-6jxqm\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.014624 4744 request.go:700] Waited for 1.910880251s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/serviceaccounts/machine-config-controller/token Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.031942 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7fd8\" (UniqueName: \"kubernetes.io/projected/dad9df42-c989-4fdf-8f19-ed7873979522-kube-api-access-j7fd8\") pod \"machine-config-controller-84d6567774-wc4mg\" (UID: \"dad9df42-c989-4fdf-8f19-ed7873979522\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.053280 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhsds\" (UniqueName: \"kubernetes.io/projected/f4865f7c-50e4-4159-8c57-42c69bd4efee-kube-api-access-jhsds\") pod \"etcd-operator-b45778765-ksbgj\" (UID: \"f4865f7c-50e4-4159-8c57-42c69bd4efee\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.069095 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jqnn\" (UniqueName: \"kubernetes.io/projected/d61086b0-b78f-4239-8d9a-49aba28c5820-kube-api-access-6jqnn\") pod \"console-f9d7485db-v5snt\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.087122 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a0c1475e-1086-40fc-800e-91f5156c92f9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2kpgr\" (UID: \"a0c1475e-1086-40fc-800e-91f5156c92f9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.110306 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f62k2\" (UniqueName: \"kubernetes.io/projected/37928f66-a94f-4241-b984-e0729ee6dbf0-kube-api-access-f62k2\") pod \"openshift-config-operator-7777fb866f-jf2dh\" (UID: \"37928f66-a94f-4241-b984-e0729ee6dbf0\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.140862 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdczj\" (UniqueName: \"kubernetes.io/projected/b550c60f-b9e0-44a5-938f-de24afa6c123-kube-api-access-hdczj\") pod \"router-default-5444994796-946kz\" (UID: \"b550c60f-b9e0-44a5-938f-de24afa6c123\") " pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.148546 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxq7v\" (UniqueName: \"kubernetes.io/projected/6ac85db3-283c-43f9-9925-b8a0f8354181-kube-api-access-hxq7v\") pod \"downloads-7954f5f757-thbg6\" (UID: \"6ac85db3-283c-43f9-9925-b8a0f8354181\") " pod="openshift-console/downloads-7954f5f757-thbg6" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.169948 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpz9s\" (UniqueName: \"kubernetes.io/projected/a0c1475e-1086-40fc-800e-91f5156c92f9-kube-api-access-hpz9s\") pod \"cluster-image-registry-operator-dc59b4c8b-2kpgr\" (UID: \"a0c1475e-1086-40fc-800e-91f5156c92f9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.173246 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.192359 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-thbg6" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.194970 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.200250 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.210364 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6t5pf" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.214011 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.218943 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.224804 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.237132 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.243639 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.259851 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.273940 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.278934 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.282792 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.283660 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-etcd-serving-ca\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.328734 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.344469 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7-image-import-ca\") pod \"apiserver-76f77b778f-2pjgv\" (UID: \"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7\") " pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.350231 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-registry-tls\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.350275 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfzdj\" (UniqueName: \"kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-kube-api-access-sfzdj\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.350325 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cc342582-3925-4af1-b9b8-8e1333af091c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.350354 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.350396 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cc342582-3925-4af1-b9b8-8e1333af091c-registry-certificates\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.350428 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc342582-3925-4af1-b9b8-8e1333af091c-trusted-ca\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.350457 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cc342582-3925-4af1-b9b8-8e1333af091c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.350479 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-bound-sa-token\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: E1003 16:27:07.355939 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:07.855917214 +0000 UTC m=+154.135793110 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.364294 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456303 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456539 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-mountpoint-dir\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456589 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3668ee6f-1314-4aec-ac55-9b55998f3c75-node-bootstrap-token\") pod \"machine-config-server-dmld4\" (UID: \"3668ee6f-1314-4aec-ac55-9b55998f3c75\") " pod="openshift-machine-config-operator/machine-config-server-dmld4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456616 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c49hm\" (UniqueName: \"kubernetes.io/projected/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-kube-api-access-c49hm\") pod \"marketplace-operator-79b997595-9hm6h\" (UID: \"cfcd176c-f3fd-419f-92ad-f7fdbcab6270\") " pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456635 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/37e50c68-0215-4d7e-aea2-e7e85dcd9b4b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vqdgh\" (UID: \"37e50c68-0215-4d7e-aea2-e7e85dcd9b4b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456655 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbrc9\" (UniqueName: \"kubernetes.io/projected/8535c47b-978d-43cf-a561-50f78e3f927d-kube-api-access-kbrc9\") pod \"packageserver-d55dfcdfc-wzkjd\" (UID: \"8535c47b-978d-43cf-a561-50f78e3f927d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456676 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd86s\" (UniqueName: \"kubernetes.io/projected/59f447fc-64dd-4628-9a42-a821c990643a-kube-api-access-jd86s\") pod \"service-ca-9c57cc56f-dpx8q\" (UID: \"59f447fc-64dd-4628-9a42-a821c990643a\") " pod="openshift-service-ca/service-ca-9c57cc56f-dpx8q" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456693 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbbb76ba-cb04-4559-a927-00e3ad2bcea0-serving-cert\") pod \"service-ca-operator-777779d784-mhrlb\" (UID: \"cbbb76ba-cb04-4559-a927-00e3ad2bcea0\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456722 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8535c47b-978d-43cf-a561-50f78e3f927d-apiservice-cert\") pod \"packageserver-d55dfcdfc-wzkjd\" (UID: \"8535c47b-978d-43cf-a561-50f78e3f927d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456742 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-registry-tls\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456761 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62497d3a-2816-455c-824e-d908c11e71e1-config-volume\") pod \"collect-profiles-29325135-jzvmv\" (UID: \"62497d3a-2816-455c-824e-d908c11e71e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456777 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-plugins-dir\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456805 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4e224494-c68b-4f4a-a32c-b631f35ad663-images\") pod \"machine-config-operator-74547568cd-cgkrq\" (UID: \"4e224494-c68b-4f4a-a32c-b631f35ad663\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456831 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8535c47b-978d-43cf-a561-50f78e3f927d-tmpfs\") pod \"packageserver-d55dfcdfc-wzkjd\" (UID: \"8535c47b-978d-43cf-a561-50f78e3f927d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456848 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c3df59a0-c796-4915-b17b-ecfd5f0d40b1-trusted-ca\") pod \"ingress-operator-5b745b69d9-ltmsn\" (UID: \"c3df59a0-c796-4915-b17b-ecfd5f0d40b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456870 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00092b0a-d430-4d0b-a90d-3a5eb74cbb88-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cffnb\" (UID: \"00092b0a-d430-4d0b-a90d-3a5eb74cbb88\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456886 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldjld\" (UniqueName: \"kubernetes.io/projected/c41f7473-52d9-4d32-9364-ed61efb65af4-kube-api-access-ldjld\") pod \"control-plane-machine-set-operator-78cbb6b69f-qmzm9\" (UID: \"c41f7473-52d9-4d32-9364-ed61efb65af4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmzm9" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456936 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a96e3d62-e019-4387-9ee1-4ea8dd90189c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cj6h4\" (UID: \"a96e3d62-e019-4387-9ee1-4ea8dd90189c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456951 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94jr6\" (UniqueName: \"kubernetes.io/projected/37e50c68-0215-4d7e-aea2-e7e85dcd9b4b-kube-api-access-94jr6\") pod \"package-server-manager-789f6589d5-vqdgh\" (UID: \"37e50c68-0215-4d7e-aea2-e7e85dcd9b4b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.456992 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frk4g\" (UniqueName: \"kubernetes.io/projected/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-kube-api-access-frk4g\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457017 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88rmh\" (UniqueName: \"kubernetes.io/projected/71471281-c2f7-42fb-8cde-6d18b8dba543-kube-api-access-88rmh\") pod \"cluster-samples-operator-665b6dd947-94qmg\" (UID: \"71471281-c2f7-42fb-8cde-6d18b8dba543\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94qmg" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457036 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c3df59a0-c796-4915-b17b-ecfd5f0d40b1-metrics-tls\") pod \"ingress-operator-5b745b69d9-ltmsn\" (UID: \"c3df59a0-c796-4915-b17b-ecfd5f0d40b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457052 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct6ff\" (UniqueName: \"kubernetes.io/projected/e93027df-377c-47a1-b2c6-9cd42db258ab-kube-api-access-ct6ff\") pod \"catalog-operator-68c6474976-q6hdh\" (UID: \"e93027df-377c-47a1-b2c6-9cd42db258ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457067 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-socket-dir\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457089 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwh72\" (UniqueName: \"kubernetes.io/projected/62497d3a-2816-455c-824e-d908c11e71e1-kube-api-access-fwh72\") pod \"collect-profiles-29325135-jzvmv\" (UID: \"62497d3a-2816-455c-824e-d908c11e71e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457107 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55164953-87bb-4683-8afd-496bbbf6385a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w4wdx\" (UID: \"55164953-87bb-4683-8afd-496bbbf6385a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457135 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3afbab4e-d827-41c5-b3a7-65061daf2eeb-config-volume\") pod \"dns-default-qmddz\" (UID: \"3afbab4e-d827-41c5-b3a7-65061daf2eeb\") " pod="openshift-dns/dns-default-qmddz" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457174 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf69c\" (UniqueName: \"kubernetes.io/projected/a96e3d62-e019-4387-9ee1-4ea8dd90189c-kube-api-access-sf69c\") pod \"olm-operator-6b444d44fb-cj6h4\" (UID: \"a96e3d62-e019-4387-9ee1-4ea8dd90189c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457190 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nmg8\" (UniqueName: \"kubernetes.io/projected/fe526e45-7a63-4289-a861-2c10510a0268-kube-api-access-2nmg8\") pod \"dns-operator-744455d44c-dfzqw\" (UID: \"fe526e45-7a63-4289-a861-2c10510a0268\") " pod="openshift-dns-operator/dns-operator-744455d44c-dfzqw" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457215 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55164953-87bb-4683-8afd-496bbbf6385a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w4wdx\" (UID: \"55164953-87bb-4683-8afd-496bbbf6385a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457230 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3c105e35-cb55-4769-a2e3-5e22ed8478c3-trusted-ca\") pod \"console-operator-58897d9998-k2tzp\" (UID: \"3c105e35-cb55-4769-a2e3-5e22ed8478c3\") " pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457276 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c105e35-cb55-4769-a2e3-5e22ed8478c3-serving-cert\") pod \"console-operator-58897d9998-k2tzp\" (UID: \"3c105e35-cb55-4769-a2e3-5e22ed8478c3\") " pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457292 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cc342582-3925-4af1-b9b8-8e1333af091c-registry-certificates\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457330 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzsr7\" (UniqueName: \"kubernetes.io/projected/3afbab4e-d827-41c5-b3a7-65061daf2eeb-kube-api-access-fzsr7\") pod \"dns-default-qmddz\" (UID: \"3afbab4e-d827-41c5-b3a7-65061daf2eeb\") " pod="openshift-dns/dns-default-qmddz" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457397 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/59f447fc-64dd-4628-9a42-a821c990643a-signing-cabundle\") pod \"service-ca-9c57cc56f-dpx8q\" (UID: \"59f447fc-64dd-4628-9a42-a821c990643a\") " pod="openshift-service-ca/service-ca-9c57cc56f-dpx8q" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457413 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbbb76ba-cb04-4559-a927-00e3ad2bcea0-config\") pod \"service-ca-operator-777779d784-mhrlb\" (UID: \"cbbb76ba-cb04-4559-a927-00e3ad2bcea0\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457460 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5c7a90d-1be3-47c3-a945-e7b872888de4-config\") pod \"kube-controller-manager-operator-78b949d7b-5js5v\" (UID: \"a5c7a90d-1be3-47c3-a945-e7b872888de4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457476 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e93027df-377c-47a1-b2c6-9cd42db258ab-profile-collector-cert\") pod \"catalog-operator-68c6474976-q6hdh\" (UID: \"e93027df-377c-47a1-b2c6-9cd42db258ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457505 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9hm6h\" (UID: \"cfcd176c-f3fd-419f-92ad-f7fdbcab6270\") " pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457543 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbqpv\" (UniqueName: \"kubernetes.io/projected/0784e651-ae4d-417f-9353-5550c790b5a4-kube-api-access-zbqpv\") pod \"ingress-canary-s96bg\" (UID: \"0784e651-ae4d-417f-9353-5550c790b5a4\") " pod="openshift-ingress-canary/ingress-canary-s96bg" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457560 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xt4p\" (UniqueName: \"kubernetes.io/projected/b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a-kube-api-access-6xt4p\") pod \"kube-storage-version-migrator-operator-b67b599dd-rspbp\" (UID: \"b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457631 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4e224494-c68b-4f4a-a32c-b631f35ad663-auth-proxy-config\") pod \"machine-config-operator-74547568cd-cgkrq\" (UID: \"4e224494-c68b-4f4a-a32c-b631f35ad663\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457646 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3668ee6f-1314-4aec-ac55-9b55998f3c75-certs\") pod \"machine-config-server-dmld4\" (UID: \"3668ee6f-1314-4aec-ac55-9b55998f3c75\") " pod="openshift-machine-config-operator/machine-config-server-dmld4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457677 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3afbab4e-d827-41c5-b3a7-65061daf2eeb-metrics-tls\") pod \"dns-default-qmddz\" (UID: \"3afbab4e-d827-41c5-b3a7-65061daf2eeb\") " pod="openshift-dns/dns-default-qmddz" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457696 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-bound-sa-token\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457712 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8535c47b-978d-43cf-a561-50f78e3f927d-webhook-cert\") pod \"packageserver-d55dfcdfc-wzkjd\" (UID: \"8535c47b-978d-43cf-a561-50f78e3f927d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457731 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c105e35-cb55-4769-a2e3-5e22ed8478c3-config\") pod \"console-operator-58897d9998-k2tzp\" (UID: \"3c105e35-cb55-4769-a2e3-5e22ed8478c3\") " pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457757 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fe526e45-7a63-4289-a861-2c10510a0268-metrics-tls\") pod \"dns-operator-744455d44c-dfzqw\" (UID: \"fe526e45-7a63-4289-a861-2c10510a0268\") " pod="openshift-dns-operator/dns-operator-744455d44c-dfzqw" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457823 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/71471281-c2f7-42fb-8cde-6d18b8dba543-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-94qmg\" (UID: \"71471281-c2f7-42fb-8cde-6d18b8dba543\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94qmg" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457860 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fgf7\" (UniqueName: \"kubernetes.io/projected/3c105e35-cb55-4769-a2e3-5e22ed8478c3-kube-api-access-9fgf7\") pod \"console-operator-58897d9998-k2tzp\" (UID: \"3c105e35-cb55-4769-a2e3-5e22ed8478c3\") " pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457887 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rspbp\" (UID: \"b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457913 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8qk6\" (UniqueName: \"kubernetes.io/projected/3668ee6f-1314-4aec-ac55-9b55998f3c75-kube-api-access-w8qk6\") pod \"machine-config-server-dmld4\" (UID: \"3668ee6f-1314-4aec-ac55-9b55998f3c75\") " pod="openshift-machine-config-operator/machine-config-server-dmld4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457929 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00092b0a-d430-4d0b-a90d-3a5eb74cbb88-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cffnb\" (UID: \"00092b0a-d430-4d0b-a90d-3a5eb74cbb88\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457944 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3df59a0-c796-4915-b17b-ecfd5f0d40b1-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ltmsn\" (UID: \"c3df59a0-c796-4915-b17b-ecfd5f0d40b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457962 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfzdj\" (UniqueName: \"kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-kube-api-access-sfzdj\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457979 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55164953-87bb-4683-8afd-496bbbf6385a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w4wdx\" (UID: \"55164953-87bb-4683-8afd-496bbbf6385a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.457994 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-csi-data-dir\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458038 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4e224494-c68b-4f4a-a32c-b631f35ad663-proxy-tls\") pod \"machine-config-operator-74547568cd-cgkrq\" (UID: \"4e224494-c68b-4f4a-a32c-b631f35ad663\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458055 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9hm6h\" (UID: \"cfcd176c-f3fd-419f-92ad-f7fdbcab6270\") " pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458073 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/16bea17d-98bf-4807-b4b3-3802d3c83e21-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-nbsh6\" (UID: \"16bea17d-98bf-4807-b4b3-3802d3c83e21\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nbsh6" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458127 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a96e3d62-e019-4387-9ee1-4ea8dd90189c-srv-cert\") pod \"olm-operator-6b444d44fb-cj6h4\" (UID: \"a96e3d62-e019-4387-9ee1-4ea8dd90189c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458146 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/59f447fc-64dd-4628-9a42-a821c990643a-signing-key\") pod \"service-ca-9c57cc56f-dpx8q\" (UID: \"59f447fc-64dd-4628-9a42-a821c990643a\") " pod="openshift-service-ca/service-ca-9c57cc56f-dpx8q" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458164 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5c7a90d-1be3-47c3-a945-e7b872888de4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5js5v\" (UID: \"a5c7a90d-1be3-47c3-a945-e7b872888de4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458181 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c41f7473-52d9-4d32-9364-ed61efb65af4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qmzm9\" (UID: \"c41f7473-52d9-4d32-9364-ed61efb65af4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmzm9" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458201 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9csbs\" (UniqueName: \"kubernetes.io/projected/cbbb76ba-cb04-4559-a927-00e3ad2bcea0-kube-api-access-9csbs\") pod \"service-ca-operator-777779d784-mhrlb\" (UID: \"cbbb76ba-cb04-4559-a927-00e3ad2bcea0\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458292 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cc342582-3925-4af1-b9b8-8e1333af091c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458320 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvc7j\" (UniqueName: \"kubernetes.io/projected/16bea17d-98bf-4807-b4b3-3802d3c83e21-kube-api-access-rvc7j\") pod \"multus-admission-controller-857f4d67dd-nbsh6\" (UID: \"16bea17d-98bf-4807-b4b3-3802d3c83e21\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nbsh6" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458348 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a5c7a90d-1be3-47c3-a945-e7b872888de4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5js5v\" (UID: \"a5c7a90d-1be3-47c3-a945-e7b872888de4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458372 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4pns\" (UniqueName: \"kubernetes.io/projected/c3df59a0-c796-4915-b17b-ecfd5f0d40b1-kube-api-access-v4pns\") pod \"ingress-operator-5b745b69d9-ltmsn\" (UID: \"c3df59a0-c796-4915-b17b-ecfd5f0d40b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458388 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rspbp\" (UID: \"b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458448 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l55x\" (UniqueName: \"kubernetes.io/projected/4e224494-c68b-4f4a-a32c-b631f35ad663-kube-api-access-9l55x\") pod \"machine-config-operator-74547568cd-cgkrq\" (UID: \"4e224494-c68b-4f4a-a32c-b631f35ad663\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458484 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e93027df-377c-47a1-b2c6-9cd42db258ab-srv-cert\") pod \"catalog-operator-68c6474976-q6hdh\" (UID: \"e93027df-377c-47a1-b2c6-9cd42db258ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458527 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k7cl\" (UniqueName: \"kubernetes.io/projected/00092b0a-d430-4d0b-a90d-3a5eb74cbb88-kube-api-access-6k7cl\") pod \"openshift-controller-manager-operator-756b6f6bc6-cffnb\" (UID: \"00092b0a-d430-4d0b-a90d-3a5eb74cbb88\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458545 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-registration-dir\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458576 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62497d3a-2816-455c-824e-d908c11e71e1-secret-volume\") pod \"collect-profiles-29325135-jzvmv\" (UID: \"62497d3a-2816-455c-824e-d908c11e71e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458621 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc342582-3925-4af1-b9b8-8e1333af091c-trusted-ca\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458640 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0784e651-ae4d-417f-9353-5550c790b5a4-cert\") pod \"ingress-canary-s96bg\" (UID: \"0784e651-ae4d-417f-9353-5550c790b5a4\") " pod="openshift-ingress-canary/ingress-canary-s96bg" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.458715 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cc342582-3925-4af1-b9b8-8e1333af091c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: E1003 16:27:07.459786 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:07.959755545 +0000 UTC m=+154.239631611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.471128 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cc342582-3925-4af1-b9b8-8e1333af091c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.474116 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-registry-tls\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.476393 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cc342582-3925-4af1-b9b8-8e1333af091c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.480172 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc342582-3925-4af1-b9b8-8e1333af091c-trusted-ca\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.480871 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cc342582-3925-4af1-b9b8-8e1333af091c-registry-certificates\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: W1003 16:27:07.482145 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb550c60f_b9e0_44a5_938f_de24afa6c123.slice/crio-25f4941b5406cb9d01cb048895a03f0637582a47b88275f2741c6ed7032d67a5 WatchSource:0}: Error finding container 25f4941b5406cb9d01cb048895a03f0637582a47b88275f2741c6ed7032d67a5: Status 404 returned error can't find the container with id 25f4941b5406cb9d01cb048895a03f0637582a47b88275f2741c6ed7032d67a5 Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.489885 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-thbg6"] Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.497601 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-bound-sa-token\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.517570 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfzdj\" (UniqueName: \"kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-kube-api-access-sfzdj\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.523887 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560448 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8535c47b-978d-43cf-a561-50f78e3f927d-tmpfs\") pod \"packageserver-d55dfcdfc-wzkjd\" (UID: \"8535c47b-978d-43cf-a561-50f78e3f927d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560489 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c3df59a0-c796-4915-b17b-ecfd5f0d40b1-trusted-ca\") pod \"ingress-operator-5b745b69d9-ltmsn\" (UID: \"c3df59a0-c796-4915-b17b-ecfd5f0d40b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560539 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00092b0a-d430-4d0b-a90d-3a5eb74cbb88-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cffnb\" (UID: \"00092b0a-d430-4d0b-a90d-3a5eb74cbb88\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560560 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldjld\" (UniqueName: \"kubernetes.io/projected/c41f7473-52d9-4d32-9364-ed61efb65af4-kube-api-access-ldjld\") pod \"control-plane-machine-set-operator-78cbb6b69f-qmzm9\" (UID: \"c41f7473-52d9-4d32-9364-ed61efb65af4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmzm9" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560582 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a96e3d62-e019-4387-9ee1-4ea8dd90189c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cj6h4\" (UID: \"a96e3d62-e019-4387-9ee1-4ea8dd90189c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560597 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94jr6\" (UniqueName: \"kubernetes.io/projected/37e50c68-0215-4d7e-aea2-e7e85dcd9b4b-kube-api-access-94jr6\") pod \"package-server-manager-789f6589d5-vqdgh\" (UID: \"37e50c68-0215-4d7e-aea2-e7e85dcd9b4b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560617 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frk4g\" (UniqueName: \"kubernetes.io/projected/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-kube-api-access-frk4g\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560632 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88rmh\" (UniqueName: \"kubernetes.io/projected/71471281-c2f7-42fb-8cde-6d18b8dba543-kube-api-access-88rmh\") pod \"cluster-samples-operator-665b6dd947-94qmg\" (UID: \"71471281-c2f7-42fb-8cde-6d18b8dba543\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94qmg" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560654 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c3df59a0-c796-4915-b17b-ecfd5f0d40b1-metrics-tls\") pod \"ingress-operator-5b745b69d9-ltmsn\" (UID: \"c3df59a0-c796-4915-b17b-ecfd5f0d40b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560669 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct6ff\" (UniqueName: \"kubernetes.io/projected/e93027df-377c-47a1-b2c6-9cd42db258ab-kube-api-access-ct6ff\") pod \"catalog-operator-68c6474976-q6hdh\" (UID: \"e93027df-377c-47a1-b2c6-9cd42db258ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560689 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-socket-dir\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560705 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwh72\" (UniqueName: \"kubernetes.io/projected/62497d3a-2816-455c-824e-d908c11e71e1-kube-api-access-fwh72\") pod \"collect-profiles-29325135-jzvmv\" (UID: \"62497d3a-2816-455c-824e-d908c11e71e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560725 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55164953-87bb-4683-8afd-496bbbf6385a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w4wdx\" (UID: \"55164953-87bb-4683-8afd-496bbbf6385a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560742 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3afbab4e-d827-41c5-b3a7-65061daf2eeb-config-volume\") pod \"dns-default-qmddz\" (UID: \"3afbab4e-d827-41c5-b3a7-65061daf2eeb\") " pod="openshift-dns/dns-default-qmddz" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560758 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf69c\" (UniqueName: \"kubernetes.io/projected/a96e3d62-e019-4387-9ee1-4ea8dd90189c-kube-api-access-sf69c\") pod \"olm-operator-6b444d44fb-cj6h4\" (UID: \"a96e3d62-e019-4387-9ee1-4ea8dd90189c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560778 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nmg8\" (UniqueName: \"kubernetes.io/projected/fe526e45-7a63-4289-a861-2c10510a0268-kube-api-access-2nmg8\") pod \"dns-operator-744455d44c-dfzqw\" (UID: \"fe526e45-7a63-4289-a861-2c10510a0268\") " pod="openshift-dns-operator/dns-operator-744455d44c-dfzqw" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560796 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55164953-87bb-4683-8afd-496bbbf6385a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w4wdx\" (UID: \"55164953-87bb-4683-8afd-496bbbf6385a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560812 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3c105e35-cb55-4769-a2e3-5e22ed8478c3-trusted-ca\") pod \"console-operator-58897d9998-k2tzp\" (UID: \"3c105e35-cb55-4769-a2e3-5e22ed8478c3\") " pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560835 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560856 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c105e35-cb55-4769-a2e3-5e22ed8478c3-serving-cert\") pod \"console-operator-58897d9998-k2tzp\" (UID: \"3c105e35-cb55-4769-a2e3-5e22ed8478c3\") " pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560875 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzsr7\" (UniqueName: \"kubernetes.io/projected/3afbab4e-d827-41c5-b3a7-65061daf2eeb-kube-api-access-fzsr7\") pod \"dns-default-qmddz\" (UID: \"3afbab4e-d827-41c5-b3a7-65061daf2eeb\") " pod="openshift-dns/dns-default-qmddz" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560891 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/59f447fc-64dd-4628-9a42-a821c990643a-signing-cabundle\") pod \"service-ca-9c57cc56f-dpx8q\" (UID: \"59f447fc-64dd-4628-9a42-a821c990643a\") " pod="openshift-service-ca/service-ca-9c57cc56f-dpx8q" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560906 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbbb76ba-cb04-4559-a927-00e3ad2bcea0-config\") pod \"service-ca-operator-777779d784-mhrlb\" (UID: \"cbbb76ba-cb04-4559-a927-00e3ad2bcea0\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560922 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5c7a90d-1be3-47c3-a945-e7b872888de4-config\") pod \"kube-controller-manager-operator-78b949d7b-5js5v\" (UID: \"a5c7a90d-1be3-47c3-a945-e7b872888de4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560943 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e93027df-377c-47a1-b2c6-9cd42db258ab-profile-collector-cert\") pod \"catalog-operator-68c6474976-q6hdh\" (UID: \"e93027df-377c-47a1-b2c6-9cd42db258ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560959 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9hm6h\" (UID: \"cfcd176c-f3fd-419f-92ad-f7fdbcab6270\") " pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560976 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xt4p\" (UniqueName: \"kubernetes.io/projected/b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a-kube-api-access-6xt4p\") pod \"kube-storage-version-migrator-operator-b67b599dd-rspbp\" (UID: \"b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.560998 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbqpv\" (UniqueName: \"kubernetes.io/projected/0784e651-ae4d-417f-9353-5550c790b5a4-kube-api-access-zbqpv\") pod \"ingress-canary-s96bg\" (UID: \"0784e651-ae4d-417f-9353-5550c790b5a4\") " pod="openshift-ingress-canary/ingress-canary-s96bg" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561019 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4e224494-c68b-4f4a-a32c-b631f35ad663-auth-proxy-config\") pod \"machine-config-operator-74547568cd-cgkrq\" (UID: \"4e224494-c68b-4f4a-a32c-b631f35ad663\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561035 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3668ee6f-1314-4aec-ac55-9b55998f3c75-certs\") pod \"machine-config-server-dmld4\" (UID: \"3668ee6f-1314-4aec-ac55-9b55998f3c75\") " pod="openshift-machine-config-operator/machine-config-server-dmld4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561051 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3afbab4e-d827-41c5-b3a7-65061daf2eeb-metrics-tls\") pod \"dns-default-qmddz\" (UID: \"3afbab4e-d827-41c5-b3a7-65061daf2eeb\") " pod="openshift-dns/dns-default-qmddz" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561069 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8535c47b-978d-43cf-a561-50f78e3f927d-webhook-cert\") pod \"packageserver-d55dfcdfc-wzkjd\" (UID: \"8535c47b-978d-43cf-a561-50f78e3f927d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561085 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c105e35-cb55-4769-a2e3-5e22ed8478c3-config\") pod \"console-operator-58897d9998-k2tzp\" (UID: \"3c105e35-cb55-4769-a2e3-5e22ed8478c3\") " pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561101 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fe526e45-7a63-4289-a861-2c10510a0268-metrics-tls\") pod \"dns-operator-744455d44c-dfzqw\" (UID: \"fe526e45-7a63-4289-a861-2c10510a0268\") " pod="openshift-dns-operator/dns-operator-744455d44c-dfzqw" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561121 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/71471281-c2f7-42fb-8cde-6d18b8dba543-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-94qmg\" (UID: \"71471281-c2f7-42fb-8cde-6d18b8dba543\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94qmg" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561141 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fgf7\" (UniqueName: \"kubernetes.io/projected/3c105e35-cb55-4769-a2e3-5e22ed8478c3-kube-api-access-9fgf7\") pod \"console-operator-58897d9998-k2tzp\" (UID: \"3c105e35-cb55-4769-a2e3-5e22ed8478c3\") " pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561157 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rspbp\" (UID: \"b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561176 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8qk6\" (UniqueName: \"kubernetes.io/projected/3668ee6f-1314-4aec-ac55-9b55998f3c75-kube-api-access-w8qk6\") pod \"machine-config-server-dmld4\" (UID: \"3668ee6f-1314-4aec-ac55-9b55998f3c75\") " pod="openshift-machine-config-operator/machine-config-server-dmld4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561196 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00092b0a-d430-4d0b-a90d-3a5eb74cbb88-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cffnb\" (UID: \"00092b0a-d430-4d0b-a90d-3a5eb74cbb88\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561214 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3df59a0-c796-4915-b17b-ecfd5f0d40b1-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ltmsn\" (UID: \"c3df59a0-c796-4915-b17b-ecfd5f0d40b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561230 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55164953-87bb-4683-8afd-496bbbf6385a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w4wdx\" (UID: \"55164953-87bb-4683-8afd-496bbbf6385a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561247 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-csi-data-dir\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561264 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4e224494-c68b-4f4a-a32c-b631f35ad663-proxy-tls\") pod \"machine-config-operator-74547568cd-cgkrq\" (UID: \"4e224494-c68b-4f4a-a32c-b631f35ad663\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561280 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9hm6h\" (UID: \"cfcd176c-f3fd-419f-92ad-f7fdbcab6270\") " pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561297 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/16bea17d-98bf-4807-b4b3-3802d3c83e21-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-nbsh6\" (UID: \"16bea17d-98bf-4807-b4b3-3802d3c83e21\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nbsh6" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561317 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a96e3d62-e019-4387-9ee1-4ea8dd90189c-srv-cert\") pod \"olm-operator-6b444d44fb-cj6h4\" (UID: \"a96e3d62-e019-4387-9ee1-4ea8dd90189c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561332 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/59f447fc-64dd-4628-9a42-a821c990643a-signing-key\") pod \"service-ca-9c57cc56f-dpx8q\" (UID: \"59f447fc-64dd-4628-9a42-a821c990643a\") " pod="openshift-service-ca/service-ca-9c57cc56f-dpx8q" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561349 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5c7a90d-1be3-47c3-a945-e7b872888de4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5js5v\" (UID: \"a5c7a90d-1be3-47c3-a945-e7b872888de4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561365 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c41f7473-52d9-4d32-9364-ed61efb65af4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qmzm9\" (UID: \"c41f7473-52d9-4d32-9364-ed61efb65af4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmzm9" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561383 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9csbs\" (UniqueName: \"kubernetes.io/projected/cbbb76ba-cb04-4559-a927-00e3ad2bcea0-kube-api-access-9csbs\") pod \"service-ca-operator-777779d784-mhrlb\" (UID: \"cbbb76ba-cb04-4559-a927-00e3ad2bcea0\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561413 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvc7j\" (UniqueName: \"kubernetes.io/projected/16bea17d-98bf-4807-b4b3-3802d3c83e21-kube-api-access-rvc7j\") pod \"multus-admission-controller-857f4d67dd-nbsh6\" (UID: \"16bea17d-98bf-4807-b4b3-3802d3c83e21\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nbsh6" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561431 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a5c7a90d-1be3-47c3-a945-e7b872888de4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5js5v\" (UID: \"a5c7a90d-1be3-47c3-a945-e7b872888de4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561447 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4pns\" (UniqueName: \"kubernetes.io/projected/c3df59a0-c796-4915-b17b-ecfd5f0d40b1-kube-api-access-v4pns\") pod \"ingress-operator-5b745b69d9-ltmsn\" (UID: \"c3df59a0-c796-4915-b17b-ecfd5f0d40b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561466 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rspbp\" (UID: \"b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561483 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l55x\" (UniqueName: \"kubernetes.io/projected/4e224494-c68b-4f4a-a32c-b631f35ad663-kube-api-access-9l55x\") pod \"machine-config-operator-74547568cd-cgkrq\" (UID: \"4e224494-c68b-4f4a-a32c-b631f35ad663\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561532 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e93027df-377c-47a1-b2c6-9cd42db258ab-srv-cert\") pod \"catalog-operator-68c6474976-q6hdh\" (UID: \"e93027df-377c-47a1-b2c6-9cd42db258ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561550 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k7cl\" (UniqueName: \"kubernetes.io/projected/00092b0a-d430-4d0b-a90d-3a5eb74cbb88-kube-api-access-6k7cl\") pod \"openshift-controller-manager-operator-756b6f6bc6-cffnb\" (UID: \"00092b0a-d430-4d0b-a90d-3a5eb74cbb88\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561567 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-registration-dir\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561584 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62497d3a-2816-455c-824e-d908c11e71e1-secret-volume\") pod \"collect-profiles-29325135-jzvmv\" (UID: \"62497d3a-2816-455c-824e-d908c11e71e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561601 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0784e651-ae4d-417f-9353-5550c790b5a4-cert\") pod \"ingress-canary-s96bg\" (UID: \"0784e651-ae4d-417f-9353-5550c790b5a4\") " pod="openshift-ingress-canary/ingress-canary-s96bg" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561601 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/8535c47b-978d-43cf-a561-50f78e3f927d-tmpfs\") pod \"packageserver-d55dfcdfc-wzkjd\" (UID: \"8535c47b-978d-43cf-a561-50f78e3f927d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561619 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-mountpoint-dir\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561685 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-mountpoint-dir\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561718 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3668ee6f-1314-4aec-ac55-9b55998f3c75-node-bootstrap-token\") pod \"machine-config-server-dmld4\" (UID: \"3668ee6f-1314-4aec-ac55-9b55998f3c75\") " pod="openshift-machine-config-operator/machine-config-server-dmld4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561751 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c49hm\" (UniqueName: \"kubernetes.io/projected/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-kube-api-access-c49hm\") pod \"marketplace-operator-79b997595-9hm6h\" (UID: \"cfcd176c-f3fd-419f-92ad-f7fdbcab6270\") " pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561779 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/37e50c68-0215-4d7e-aea2-e7e85dcd9b4b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vqdgh\" (UID: \"37e50c68-0215-4d7e-aea2-e7e85dcd9b4b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561807 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbrc9\" (UniqueName: \"kubernetes.io/projected/8535c47b-978d-43cf-a561-50f78e3f927d-kube-api-access-kbrc9\") pod \"packageserver-d55dfcdfc-wzkjd\" (UID: \"8535c47b-978d-43cf-a561-50f78e3f927d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561826 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd86s\" (UniqueName: \"kubernetes.io/projected/59f447fc-64dd-4628-9a42-a821c990643a-kube-api-access-jd86s\") pod \"service-ca-9c57cc56f-dpx8q\" (UID: \"59f447fc-64dd-4628-9a42-a821c990643a\") " pod="openshift-service-ca/service-ca-9c57cc56f-dpx8q" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561847 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbbb76ba-cb04-4559-a927-00e3ad2bcea0-serving-cert\") pod \"service-ca-operator-777779d784-mhrlb\" (UID: \"cbbb76ba-cb04-4559-a927-00e3ad2bcea0\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561875 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8535c47b-978d-43cf-a561-50f78e3f927d-apiservice-cert\") pod \"packageserver-d55dfcdfc-wzkjd\" (UID: \"8535c47b-978d-43cf-a561-50f78e3f927d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561893 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62497d3a-2816-455c-824e-d908c11e71e1-config-volume\") pod \"collect-profiles-29325135-jzvmv\" (UID: \"62497d3a-2816-455c-824e-d908c11e71e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561909 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-plugins-dir\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.561933 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4e224494-c68b-4f4a-a32c-b631f35ad663-images\") pod \"machine-config-operator-74547568cd-cgkrq\" (UID: \"4e224494-c68b-4f4a-a32c-b631f35ad663\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.562314 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3afbab4e-d827-41c5-b3a7-65061daf2eeb-config-volume\") pod \"dns-default-qmddz\" (UID: \"3afbab4e-d827-41c5-b3a7-65061daf2eeb\") " pod="openshift-dns/dns-default-qmddz" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.562533 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-socket-dir\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.562673 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4e224494-c68b-4f4a-a32c-b631f35ad663-images\") pod \"machine-config-operator-74547568cd-cgkrq\" (UID: \"4e224494-c68b-4f4a-a32c-b631f35ad663\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.563618 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c3df59a0-c796-4915-b17b-ecfd5f0d40b1-trusted-ca\") pod \"ingress-operator-5b745b69d9-ltmsn\" (UID: \"c3df59a0-c796-4915-b17b-ecfd5f0d40b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.564078 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00092b0a-d430-4d0b-a90d-3a5eb74cbb88-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cffnb\" (UID: \"00092b0a-d430-4d0b-a90d-3a5eb74cbb88\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.571081 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55164953-87bb-4683-8afd-496bbbf6385a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w4wdx\" (UID: \"55164953-87bb-4683-8afd-496bbbf6385a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.571590 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a96e3d62-e019-4387-9ee1-4ea8dd90189c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cj6h4\" (UID: \"a96e3d62-e019-4387-9ee1-4ea8dd90189c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.571871 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55164953-87bb-4683-8afd-496bbbf6385a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w4wdx\" (UID: \"55164953-87bb-4683-8afd-496bbbf6385a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.571968 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-csi-data-dir\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: E1003 16:27:07.575376 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:08.075357037 +0000 UTC m=+154.355232933 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.576024 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rspbp\" (UID: \"b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.576079 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3c105e35-cb55-4769-a2e3-5e22ed8478c3-trusted-ca\") pod \"console-operator-58897d9998-k2tzp\" (UID: \"3c105e35-cb55-4769-a2e3-5e22ed8478c3\") " pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.576938 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00092b0a-d430-4d0b-a90d-3a5eb74cbb88-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cffnb\" (UID: \"00092b0a-d430-4d0b-a90d-3a5eb74cbb88\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.577032 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-registration-dir\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.582423 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8535c47b-978d-43cf-a561-50f78e3f927d-apiservice-cert\") pod \"packageserver-d55dfcdfc-wzkjd\" (UID: \"8535c47b-978d-43cf-a561-50f78e3f927d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.584178 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62497d3a-2816-455c-824e-d908c11e71e1-config-volume\") pod \"collect-profiles-29325135-jzvmv\" (UID: \"62497d3a-2816-455c-824e-d908c11e71e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.584237 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-plugins-dir\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.590999 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c41f7473-52d9-4d32-9364-ed61efb65af4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qmzm9\" (UID: \"c41f7473-52d9-4d32-9364-ed61efb65af4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmzm9" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.595087 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62497d3a-2816-455c-824e-d908c11e71e1-secret-volume\") pod \"collect-profiles-29325135-jzvmv\" (UID: \"62497d3a-2816-455c-824e-d908c11e71e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.596400 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/59f447fc-64dd-4628-9a42-a821c990643a-signing-cabundle\") pod \"service-ca-9c57cc56f-dpx8q\" (UID: \"59f447fc-64dd-4628-9a42-a821c990643a\") " pod="openshift-service-ca/service-ca-9c57cc56f-dpx8q" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.600050 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/16bea17d-98bf-4807-b4b3-3802d3c83e21-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-nbsh6\" (UID: \"16bea17d-98bf-4807-b4b3-3802d3c83e21\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nbsh6" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.600156 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0784e651-ae4d-417f-9353-5550c790b5a4-cert\") pod \"ingress-canary-s96bg\" (UID: \"0784e651-ae4d-417f-9353-5550c790b5a4\") " pod="openshift-ingress-canary/ingress-canary-s96bg" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.600419 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5c7a90d-1be3-47c3-a945-e7b872888de4-config\") pod \"kube-controller-manager-operator-78b949d7b-5js5v\" (UID: \"a5c7a90d-1be3-47c3-a945-e7b872888de4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.600511 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/59f447fc-64dd-4628-9a42-a821c990643a-signing-key\") pod \"service-ca-9c57cc56f-dpx8q\" (UID: \"59f447fc-64dd-4628-9a42-a821c990643a\") " pod="openshift-service-ca/service-ca-9c57cc56f-dpx8q" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.601663 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9hm6h\" (UID: \"cfcd176c-f3fd-419f-92ad-f7fdbcab6270\") " pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.603431 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbbb76ba-cb04-4559-a927-00e3ad2bcea0-config\") pod \"service-ca-operator-777779d784-mhrlb\" (UID: \"cbbb76ba-cb04-4559-a927-00e3ad2bcea0\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.603806 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e93027df-377c-47a1-b2c6-9cd42db258ab-profile-collector-cert\") pod \"catalog-operator-68c6474976-q6hdh\" (UID: \"e93027df-377c-47a1-b2c6-9cd42db258ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.604550 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c105e35-cb55-4769-a2e3-5e22ed8478c3-config\") pod \"console-operator-58897d9998-k2tzp\" (UID: \"3c105e35-cb55-4769-a2e3-5e22ed8478c3\") " pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.604696 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4e224494-c68b-4f4a-a32c-b631f35ad663-auth-proxy-config\") pod \"machine-config-operator-74547568cd-cgkrq\" (UID: \"4e224494-c68b-4f4a-a32c-b631f35ad663\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.605140 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rspbp\" (UID: \"b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.605770 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3668ee6f-1314-4aec-ac55-9b55998f3c75-certs\") pod \"machine-config-server-dmld4\" (UID: \"3668ee6f-1314-4aec-ac55-9b55998f3c75\") " pod="openshift-machine-config-operator/machine-config-server-dmld4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.606041 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbbb76ba-cb04-4559-a927-00e3ad2bcea0-serving-cert\") pod \"service-ca-operator-777779d784-mhrlb\" (UID: \"cbbb76ba-cb04-4559-a927-00e3ad2bcea0\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.612279 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e93027df-377c-47a1-b2c6-9cd42db258ab-srv-cert\") pod \"catalog-operator-68c6474976-q6hdh\" (UID: \"e93027df-377c-47a1-b2c6-9cd42db258ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.613967 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fe526e45-7a63-4289-a861-2c10510a0268-metrics-tls\") pod \"dns-operator-744455d44c-dfzqw\" (UID: \"fe526e45-7a63-4289-a861-2c10510a0268\") " pod="openshift-dns-operator/dns-operator-744455d44c-dfzqw" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.626119 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/71471281-c2f7-42fb-8cde-6d18b8dba543-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-94qmg\" (UID: \"71471281-c2f7-42fb-8cde-6d18b8dba543\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94qmg" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.626300 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4e224494-c68b-4f4a-a32c-b631f35ad663-proxy-tls\") pod \"machine-config-operator-74547568cd-cgkrq\" (UID: \"4e224494-c68b-4f4a-a32c-b631f35ad663\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.627037 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9hm6h\" (UID: \"cfcd176c-f3fd-419f-92ad-f7fdbcab6270\") " pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.627110 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c3df59a0-c796-4915-b17b-ecfd5f0d40b1-metrics-tls\") pod \"ingress-operator-5b745b69d9-ltmsn\" (UID: \"c3df59a0-c796-4915-b17b-ecfd5f0d40b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.627574 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c105e35-cb55-4769-a2e3-5e22ed8478c3-serving-cert\") pod \"console-operator-58897d9998-k2tzp\" (UID: \"3c105e35-cb55-4769-a2e3-5e22ed8478c3\") " pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.630031 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3afbab4e-d827-41c5-b3a7-65061daf2eeb-metrics-tls\") pod \"dns-default-qmddz\" (UID: \"3afbab4e-d827-41c5-b3a7-65061daf2eeb\") " pod="openshift-dns/dns-default-qmddz" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.631051 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5c7a90d-1be3-47c3-a945-e7b872888de4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5js5v\" (UID: \"a5c7a90d-1be3-47c3-a945-e7b872888de4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.639579 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/37e50c68-0215-4d7e-aea2-e7e85dcd9b4b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vqdgh\" (UID: \"37e50c68-0215-4d7e-aea2-e7e85dcd9b4b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.640015 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3668ee6f-1314-4aec-ac55-9b55998f3c75-node-bootstrap-token\") pod \"machine-config-server-dmld4\" (UID: \"3668ee6f-1314-4aec-ac55-9b55998f3c75\") " pod="openshift-machine-config-operator/machine-config-server-dmld4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.648351 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8535c47b-978d-43cf-a561-50f78e3f927d-webhook-cert\") pod \"packageserver-d55dfcdfc-wzkjd\" (UID: \"8535c47b-978d-43cf-a561-50f78e3f927d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.649810 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55164953-87bb-4683-8afd-496bbbf6385a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w4wdx\" (UID: \"55164953-87bb-4683-8afd-496bbbf6385a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.662078 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.663160 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:07 crc kubenswrapper[4744]: E1003 16:27:07.663744 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:08.163721464 +0000 UTC m=+154.443597350 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.665315 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a96e3d62-e019-4387-9ee1-4ea8dd90189c-srv-cert\") pod \"olm-operator-6b444d44fb-cj6h4\" (UID: \"a96e3d62-e019-4387-9ee1-4ea8dd90189c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.683181 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nmg8\" (UniqueName: \"kubernetes.io/projected/fe526e45-7a63-4289-a861-2c10510a0268-kube-api-access-2nmg8\") pod \"dns-operator-744455d44c-dfzqw\" (UID: \"fe526e45-7a63-4289-a861-2c10510a0268\") " pod="openshift-dns-operator/dns-operator-744455d44c-dfzqw" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.705232 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf69c\" (UniqueName: \"kubernetes.io/projected/a96e3d62-e019-4387-9ee1-4ea8dd90189c-kube-api-access-sf69c\") pod \"olm-operator-6b444d44fb-cj6h4\" (UID: \"a96e3d62-e019-4387-9ee1-4ea8dd90189c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.720150 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwh72\" (UniqueName: \"kubernetes.io/projected/62497d3a-2816-455c-824e-d908c11e71e1-kube-api-access-fwh72\") pod \"collect-profiles-29325135-jzvmv\" (UID: \"62497d3a-2816-455c-824e-d908c11e71e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.727455 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-946kz" event={"ID":"b550c60f-b9e0-44a5-938f-de24afa6c123","Type":"ContainerStarted","Data":"25f4941b5406cb9d01cb048895a03f0637582a47b88275f2741c6ed7032d67a5"} Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.728401 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-thbg6" event={"ID":"6ac85db3-283c-43f9-9925-b8a0f8354181","Type":"ContainerStarted","Data":"8c4f1fc28ba55e6f8ae69e99a7aa5410276d9d0bfc042c2e884e242f5ca27156"} Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.730114 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" event={"ID":"90345bca-bb79-4ea3-9cc8-f6f8bebb3180","Type":"ContainerStarted","Data":"b725e7baca29cdb8bc54252e0d63ff28e4f3848b7c60618b7ad6636bc0bc8632"} Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.738380 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" event={"ID":"e50b3ffc-b380-4fd2-92f6-35b6b1dc5bd4","Type":"ContainerStarted","Data":"68b6074bfb762fdec5179817cef4b3d23e27cd0229e20d6a182d1c6a8c5ca799"} Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.740591 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldjld\" (UniqueName: \"kubernetes.io/projected/c41f7473-52d9-4d32-9364-ed61efb65af4-kube-api-access-ldjld\") pod \"control-plane-machine-set-operator-78cbb6b69f-qmzm9\" (UID: \"c41f7473-52d9-4d32-9364-ed61efb65af4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmzm9" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.754721 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.758385 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.760442 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3df59a0-c796-4915-b17b-ecfd5f0d40b1-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ltmsn\" (UID: \"c3df59a0-c796-4915-b17b-ecfd5f0d40b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.765691 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: E1003 16:27:07.766124 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:08.266107865 +0000 UTC m=+154.545983761 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.768352 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9csbs\" (UniqueName: \"kubernetes.io/projected/cbbb76ba-cb04-4559-a927-00e3ad2bcea0-kube-api-access-9csbs\") pod \"service-ca-operator-777779d784-mhrlb\" (UID: \"cbbb76ba-cb04-4559-a927-00e3ad2bcea0\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.771173 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-t9pf9"] Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.789164 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.801404 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94jr6\" (UniqueName: \"kubernetes.io/projected/37e50c68-0215-4d7e-aea2-e7e85dcd9b4b-kube-api-access-94jr6\") pod \"package-server-manager-789f6589d5-vqdgh\" (UID: \"37e50c68-0215-4d7e-aea2-e7e85dcd9b4b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.833981 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88rmh\" (UniqueName: \"kubernetes.io/projected/71471281-c2f7-42fb-8cde-6d18b8dba543-kube-api-access-88rmh\") pod \"cluster-samples-operator-665b6dd947-94qmg\" (UID: \"71471281-c2f7-42fb-8cde-6d18b8dba543\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94qmg" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.834157 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frk4g\" (UniqueName: \"kubernetes.io/projected/36b56a51-ac0b-437b-8f4f-e42476ed6ddd-kube-api-access-frk4g\") pod \"csi-hostpathplugin-d6zzl\" (UID: \"36b56a51-ac0b-437b-8f4f-e42476ed6ddd\") " pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.851110 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvc7j\" (UniqueName: \"kubernetes.io/projected/16bea17d-98bf-4807-b4b3-3802d3c83e21-kube-api-access-rvc7j\") pod \"multus-admission-controller-857f4d67dd-nbsh6\" (UID: \"16bea17d-98bf-4807-b4b3-3802d3c83e21\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nbsh6" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.871915 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:07 crc kubenswrapper[4744]: E1003 16:27:07.872743 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:08.372725982 +0000 UTC m=+154.652601868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.888035 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6t5pf"] Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.897411 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94qmg" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.899724 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l55x\" (UniqueName: \"kubernetes.io/projected/4e224494-c68b-4f4a-a32c-b631f35ad663-kube-api-access-9l55x\") pod \"machine-config-operator-74547568cd-cgkrq\" (UID: \"4e224494-c68b-4f4a-a32c-b631f35ad663\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.902926 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.907185 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4pns\" (UniqueName: \"kubernetes.io/projected/c3df59a0-c796-4915-b17b-ecfd5f0d40b1-kube-api-access-v4pns\") pod \"ingress-operator-5b745b69d9-ltmsn\" (UID: \"c3df59a0-c796-4915-b17b-ecfd5f0d40b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.908086 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.922606 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ksbgj"] Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.925287 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a5c7a90d-1be3-47c3-a945-e7b872888de4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5js5v\" (UID: \"a5c7a90d-1be3-47c3-a945-e7b872888de4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.925721 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-dfzqw" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.946268 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmzm9" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.952922 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-nbsh6" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.958772 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct6ff\" (UniqueName: \"kubernetes.io/projected/e93027df-377c-47a1-b2c6-9cd42db258ab-kube-api-access-ct6ff\") pod \"catalog-operator-68c6474976-q6hdh\" (UID: \"e93027df-377c-47a1-b2c6-9cd42db258ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.963092 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k7cl\" (UniqueName: \"kubernetes.io/projected/00092b0a-d430-4d0b-a90d-3a5eb74cbb88-kube-api-access-6k7cl\") pod \"openshift-controller-manager-operator-756b6f6bc6-cffnb\" (UID: \"00092b0a-d430-4d0b-a90d-3a5eb74cbb88\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.973010 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:07 crc kubenswrapper[4744]: E1003 16:27:07.973445 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:08.473424226 +0000 UTC m=+154.753300122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.979339 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd86s\" (UniqueName: \"kubernetes.io/projected/59f447fc-64dd-4628-9a42-a821c990643a-kube-api-access-jd86s\") pod \"service-ca-9c57cc56f-dpx8q\" (UID: \"59f447fc-64dd-4628-9a42-a821c990643a\") " pod="openshift-service-ca/service-ca-9c57cc56f-dpx8q" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.979772 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.983337 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c49hm\" (UniqueName: \"kubernetes.io/projected/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-kube-api-access-c49hm\") pod \"marketplace-operator-79b997595-9hm6h\" (UID: \"cfcd176c-f3fd-419f-92ad-f7fdbcab6270\") " pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" Oct 03 16:27:07 crc kubenswrapper[4744]: I1003 16:27:07.985091 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.006389 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzsr7\" (UniqueName: \"kubernetes.io/projected/3afbab4e-d827-41c5-b3a7-65061daf2eeb-kube-api-access-fzsr7\") pod \"dns-default-qmddz\" (UID: \"3afbab4e-d827-41c5-b3a7-65061daf2eeb\") " pod="openshift-dns/dns-default-qmddz" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.007061 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.009309 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbrc9\" (UniqueName: \"kubernetes.io/projected/8535c47b-978d-43cf-a561-50f78e3f927d-kube-api-access-kbrc9\") pod \"packageserver-d55dfcdfc-wzkjd\" (UID: \"8535c47b-978d-43cf-a561-50f78e3f927d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.010386 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.020160 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.038384 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xt4p\" (UniqueName: \"kubernetes.io/projected/b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a-kube-api-access-6xt4p\") pod \"kube-storage-version-migrator-operator-b67b599dd-rspbp\" (UID: \"b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.051769 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.063904 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbqpv\" (UniqueName: \"kubernetes.io/projected/0784e651-ae4d-417f-9353-5550c790b5a4-kube-api-access-zbqpv\") pod \"ingress-canary-s96bg\" (UID: \"0784e651-ae4d-417f-9353-5550c790b5a4\") " pod="openshift-ingress-canary/ingress-canary-s96bg" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.068676 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.075093 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:08 crc kubenswrapper[4744]: E1003 16:27:08.075306 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:08.575272582 +0000 UTC m=+154.855148478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.075361 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:08 crc kubenswrapper[4744]: E1003 16:27:08.075852 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:08.575844988 +0000 UTC m=+154.855720884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.076711 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fgf7\" (UniqueName: \"kubernetes.io/projected/3c105e35-cb55-4769-a2e3-5e22ed8478c3-kube-api-access-9fgf7\") pod \"console-operator-58897d9998-k2tzp\" (UID: \"3c105e35-cb55-4769-a2e3-5e22ed8478c3\") " pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.077321 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-qmddz" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.085327 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-s96bg" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.116239 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8qk6\" (UniqueName: \"kubernetes.io/projected/3668ee6f-1314-4aec-ac55-9b55998f3c75-kube-api-access-w8qk6\") pod \"machine-config-server-dmld4\" (UID: \"3668ee6f-1314-4aec-ac55-9b55998f3c75\") " pod="openshift-machine-config-operator/machine-config-server-dmld4" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.180874 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:08 crc kubenswrapper[4744]: E1003 16:27:08.181416 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:08.681391855 +0000 UTC m=+154.961267751 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.194921 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.235395 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.239843 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.248125 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-v5snt"] Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.253828 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg"] Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.260516 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dpx8q" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.272873 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.283827 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:08 crc kubenswrapper[4744]: E1003 16:27:08.284208 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:08.784191877 +0000 UTC m=+155.064067773 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.343637 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6jxqm"] Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.384791 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:08 crc kubenswrapper[4744]: E1003 16:27:08.385190 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:08.885174519 +0000 UTC m=+155.165050415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.392219 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-dmld4" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.433059 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh"] Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.496580 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:08 crc kubenswrapper[4744]: E1003 16:27:08.496950 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:08.996933047 +0000 UTC m=+155.276808943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.501476 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2pjgv"] Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.502810 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7"] Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.519078 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr"] Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.562265 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94qmg"] Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.598334 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:08 crc kubenswrapper[4744]: E1003 16:27:08.598937 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:09.098875355 +0000 UTC m=+155.378751351 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.699899 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:08 crc kubenswrapper[4744]: E1003 16:27:08.700926 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:09.200901246 +0000 UTC m=+155.480777332 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.765108 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" podStartSLOduration=132.765074221 podStartE2EDuration="2m12.765074221s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:08.764322021 +0000 UTC m=+155.044197917" watchObservedRunningTime="2025-10-03 16:27:08.765074221 +0000 UTC m=+155.044950117" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.801278 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:08 crc kubenswrapper[4744]: E1003 16:27:08.802441 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:09.302403223 +0000 UTC m=+155.582279119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.855049 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6t5pf" event={"ID":"b2a7d1a9-0abe-4150-8c23-c9f592467ba5","Type":"ContainerStarted","Data":"b58693a6a8e2aed7e0384491232624b8ab28322565e88c9045b37711294412fd"} Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.859825 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" event={"ID":"37928f66-a94f-4241-b984-e0729ee6dbf0","Type":"ContainerStarted","Data":"01c1fc07f78711754c8ab898e874a989a9d75d43e3ecf52000ee5a7b6a422937"} Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.862074 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" event={"ID":"f4865f7c-50e4-4159-8c57-42c69bd4efee","Type":"ContainerStarted","Data":"657a9252e80ebfa60dfe672ceb00b7b9eb4ade5850a8ea2a26232d2d4f26bf11"} Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.890181 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-thbg6" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.896129 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-b8wvx" podStartSLOduration=133.896109406 podStartE2EDuration="2m13.896109406s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:08.866757573 +0000 UTC m=+155.146633469" watchObservedRunningTime="2025-10-03 16:27:08.896109406 +0000 UTC m=+155.175985302" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.902725 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:08 crc kubenswrapper[4744]: E1003 16:27:08.903065 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:09.403051626 +0000 UTC m=+155.682927522 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.914175 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg" event={"ID":"dad9df42-c989-4fdf-8f19-ed7873979522","Type":"ContainerStarted","Data":"cb46744107697b91665493da38e68513de9f731d0034fa40ad6f65d5b50dc735"} Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.920212 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-thbg6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.920283 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-thbg6" podUID="6ac85db3-283c-43f9-9925-b8a0f8354181" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.927208 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-946kz" event={"ID":"b550c60f-b9e0-44a5-938f-de24afa6c123","Type":"ContainerStarted","Data":"7d69ddbe0bfd47ebd502b9086a80264d585ba9663d1a52e9bb783c442ccdbb2b"} Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.928805 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" event={"ID":"4f39c88b-8d2a-4a19-921e-14d8d172ad20","Type":"ContainerStarted","Data":"1b565c39e26d076105a185b95e565027488a1d2fc456f55e024f38e20c52151b"} Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.929400 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4"] Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.930070 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" event={"ID":"4b772820-bbb4-4dec-937d-6502175fae6a","Type":"ContainerStarted","Data":"8674d133686399a33c8a8693935571347bea7f6ec968d22b4ebdcccb109327ae"} Oct 03 16:27:08 crc kubenswrapper[4744]: I1003 16:27:08.937550 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-v5snt" event={"ID":"d61086b0-b78f-4239-8d9a-49aba28c5820","Type":"ContainerStarted","Data":"ab1fb08e373c4fe5a018e7df29d9d99216c0be26b95d9127c213773e2d982cb6"} Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.007221 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:09 crc kubenswrapper[4744]: E1003 16:27:09.008435 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:09.508386997 +0000 UTC m=+155.788262893 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.067397 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-btrfh" podStartSLOduration=133.067373501 podStartE2EDuration="2m13.067373501s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:09.059531916 +0000 UTC m=+155.339407822" watchObservedRunningTime="2025-10-03 16:27:09.067373501 +0000 UTC m=+155.347249397" Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.114610 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:09 crc kubenswrapper[4744]: E1003 16:27:09.115033 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:09.615018984 +0000 UTC m=+155.894894880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.223034 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" podStartSLOduration=134.223012488 podStartE2EDuration="2m14.223012488s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:09.20445505 +0000 UTC m=+155.484330956" watchObservedRunningTime="2025-10-03 16:27:09.223012488 +0000 UTC m=+155.502888384" Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.223853 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq"] Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.234660 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:09 crc kubenswrapper[4744]: E1003 16:27:09.235009 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:09.734989356 +0000 UTC m=+156.014865252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.269582 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.307673 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx"] Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.343646 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:09 crc kubenswrapper[4744]: E1003 16:27:09.344149 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:09.84413118 +0000 UTC m=+156.124007076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.380038 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv"] Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.389159 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-nbsh6"] Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.390703 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn"] Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.420701 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmzm9"] Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.464238 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.468646 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:09 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:09 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:09 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.468885 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:09 crc kubenswrapper[4744]: E1003 16:27:09.474011 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:09.973962782 +0000 UTC m=+156.253838688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:09 crc kubenswrapper[4744]: E1003 16:27:09.475629 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:09.975584686 +0000 UTC m=+156.255460582 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.476092 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.502771 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t84ms" podStartSLOduration=135.502735229 podStartE2EDuration="2m15.502735229s" podCreationTimestamp="2025-10-03 16:24:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:09.453471111 +0000 UTC m=+155.733347007" watchObservedRunningTime="2025-10-03 16:27:09.502735229 +0000 UTC m=+155.782611125" Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.579043 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:09 crc kubenswrapper[4744]: E1003 16:27:09.579599 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:10.07956085 +0000 UTC m=+156.359436776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.657207 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" podStartSLOduration=133.657184354 podStartE2EDuration="2m13.657184354s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:09.65448533 +0000 UTC m=+155.934361226" watchObservedRunningTime="2025-10-03 16:27:09.657184354 +0000 UTC m=+155.937060240" Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.683697 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:09 crc kubenswrapper[4744]: E1003 16:27:09.684121 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:10.18410711 +0000 UTC m=+156.463983006 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.737439 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-thbg6" podStartSLOduration=134.737404128 podStartE2EDuration="2m14.737404128s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:09.730660514 +0000 UTC m=+156.010536410" watchObservedRunningTime="2025-10-03 16:27:09.737404128 +0000 UTC m=+156.017280024" Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.780271 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-946kz" podStartSLOduration=133.78024511 podStartE2EDuration="2m13.78024511s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:09.777081744 +0000 UTC m=+156.056957640" watchObservedRunningTime="2025-10-03 16:27:09.78024511 +0000 UTC m=+156.060121006" Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.784379 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:09 crc kubenswrapper[4744]: E1003 16:27:09.784742 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:10.284716212 +0000 UTC m=+156.564592108 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.892669 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:09 crc kubenswrapper[4744]: E1003 16:27:09.893165 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:10.393147679 +0000 UTC m=+156.673023575 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:09 crc kubenswrapper[4744]: I1003 16:27:09.995611 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:09 crc kubenswrapper[4744]: E1003 16:27:09.996513 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:10.496474005 +0000 UTC m=+156.776349901 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.049974 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" event={"ID":"a96e3d62-e019-4387-9ee1-4ea8dd90189c","Type":"ContainerStarted","Data":"37bc7502c4c593a6995df74319f8c48b7c770d45fdcc95520ad86ef5d997f693"} Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.075909 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" event={"ID":"4e224494-c68b-4f4a-a32c-b631f35ad663","Type":"ContainerStarted","Data":"96d21ad3563ff4dd340f77261e7e983f7e7141c81da5887a86581e7582290f99"} Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.084304 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh"] Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.088122 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-dfzqw"] Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.098529 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:10 crc kubenswrapper[4744]: E1003 16:27:10.099220 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:10.599203865 +0000 UTC m=+156.879079761 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.104112 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" event={"ID":"62497d3a-2816-455c-824e-d908c11e71e1","Type":"ContainerStarted","Data":"f4e8b264982619990bf78dccca62adade613a46d3ef8d825f82d5a3d4779ae2a"} Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.117211 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" event={"ID":"a0c1475e-1086-40fc-800e-91f5156c92f9","Type":"ContainerStarted","Data":"6635d9c8bbebcc225dde35701e95e1cf35af746a0711491ead6ed842aa7412a1"} Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.121839 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94qmg" event={"ID":"71471281-c2f7-42fb-8cde-6d18b8dba543","Type":"ContainerStarted","Data":"4cd615b00bc51ada612bfdcbd782cfa6e253276c4ba35fe8db0b98afc96e68a8"} Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.138378 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6t5pf" event={"ID":"b2a7d1a9-0abe-4150-8c23-c9f592467ba5","Type":"ContainerStarted","Data":"bbd24dfb5c9596470837d3f4bc4fd7e2534dd880a08752f58b5c38748d16f083"} Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.166078 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-nbsh6" event={"ID":"16bea17d-98bf-4807-b4b3-3802d3c83e21","Type":"ContainerStarted","Data":"fb7e3d93a3f090f00b93f1b9349213f3b5334d781ed1bcfa15e1eae3503e4866"} Oct 03 16:27:10 crc kubenswrapper[4744]: W1003 16:27:10.173776 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe526e45_7a63_4289_a861_2c10510a0268.slice/crio-2c4afe3df424570307a41bd157463b8cf0cf347a824eb338b7e492d336adc27c WatchSource:0}: Error finding container 2c4afe3df424570307a41bd157463b8cf0cf347a824eb338b7e492d336adc27c: Status 404 returned error can't find the container with id 2c4afe3df424570307a41bd157463b8cf0cf347a824eb338b7e492d336adc27c Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.185518 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7" event={"ID":"a0b52a53-304a-4664-b8d7-9434f969dac3","Type":"ContainerStarted","Data":"2ebd2044a5c6d43fa3d53c3594288136a82e41d14f55e55c8e96b63493f2e41c"} Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.199876 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:10 crc kubenswrapper[4744]: E1003 16:27:10.200121 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:10.700079485 +0000 UTC m=+156.979955391 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.200462 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:10 crc kubenswrapper[4744]: E1003 16:27:10.201014 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:10.70100493 +0000 UTC m=+156.980880826 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.213288 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-thbg6" event={"ID":"6ac85db3-283c-43f9-9925-b8a0f8354181","Type":"ContainerStarted","Data":"75d5b59ccdee14ad8ec29fb6a326b6dbeb4e637f6b11cd8cf0cf210258d4be15"} Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.215576 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" event={"ID":"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7","Type":"ContainerStarted","Data":"e51b7fa40ca1289fda2ad2b01e7c82244f0f2c21b8245bb71754489bfca2dd47"} Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.217780 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" event={"ID":"c3df59a0-c796-4915-b17b-ecfd5f0d40b1","Type":"ContainerStarted","Data":"df55ba4c2de7c396717d02f8a7883b8e6a8e6135cd2abee059927d1930ae9db3"} Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.218038 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-thbg6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.218088 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-thbg6" podUID="6ac85db3-283c-43f9-9925-b8a0f8354181" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.237585 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-dmld4" event={"ID":"3668ee6f-1314-4aec-ac55-9b55998f3c75","Type":"ContainerStarted","Data":"62dab15b1df54d076316f6fd2080b1fd5688d174115b0c99f819d695682bef75"} Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.239545 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmzm9" event={"ID":"c41f7473-52d9-4d32-9364-ed61efb65af4","Type":"ContainerStarted","Data":"0dd4f779ee46ecdf178b606c895304a406bbe1dd71a09991a5002d91e9a4a16e"} Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.242350 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx" event={"ID":"55164953-87bb-4683-8afd-496bbbf6385a","Type":"ContainerStarted","Data":"6580ca40fba4d605b8ca895cd10ed06696a2cd3d1e2d341a931c25de8a33f2c5"} Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.243613 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" event={"ID":"4b772820-bbb4-4dec-937d-6502175fae6a","Type":"ContainerStarted","Data":"f275c2282f01bd452a4397ca4fe340e6abe8f717fec1707010c0b224b95a26c5"} Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.286044 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:10 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:10 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:10 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.286553 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.304662 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:10 crc kubenswrapper[4744]: E1003 16:27:10.306404 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:10.806382262 +0000 UTC m=+157.086258158 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.324134 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-t9pf9" podStartSLOduration=135.324106857 podStartE2EDuration="2m15.324106857s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:10.317791254 +0000 UTC m=+156.597667150" watchObservedRunningTime="2025-10-03 16:27:10.324106857 +0000 UTC m=+156.603982753" Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.374750 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh"] Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.378747 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd"] Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.388968 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9hm6h"] Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.402816 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-k2tzp"] Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.407821 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:10 crc kubenswrapper[4744]: E1003 16:27:10.408533 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:10.908517806 +0000 UTC m=+157.188393702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.414302 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v"] Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.422752 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-s96bg"] Oct 03 16:27:10 crc kubenswrapper[4744]: W1003 16:27:10.453758 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfcd176c_f3fd_419f_92ad_f7fdbcab6270.slice/crio-286a60c0444c4cf67a80c02e6137b130e8c68a28048b05284a094712b3b7cfe1 WatchSource:0}: Error finding container 286a60c0444c4cf67a80c02e6137b130e8c68a28048b05284a094712b3b7cfe1: Status 404 returned error can't find the container with id 286a60c0444c4cf67a80c02e6137b130e8c68a28048b05284a094712b3b7cfe1 Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.478929 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dpx8q"] Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.502406 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb"] Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.504949 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb"] Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.506567 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-qmddz"] Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.512591 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:10 crc kubenswrapper[4744]: E1003 16:27:10.513146 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:11.013115847 +0000 UTC m=+157.292991743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.518167 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-d6zzl"] Oct 03 16:27:10 crc kubenswrapper[4744]: W1003 16:27:10.536883 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00092b0a_d430_4d0b_a90d_3a5eb74cbb88.slice/crio-db9165efacb55f8061214ab7349927d295b014de051a2664eeafd770f16daf17 WatchSource:0}: Error finding container db9165efacb55f8061214ab7349927d295b014de051a2664eeafd770f16daf17: Status 404 returned error can't find the container with id db9165efacb55f8061214ab7349927d295b014de051a2664eeafd770f16daf17 Oct 03 16:27:10 crc kubenswrapper[4744]: W1003 16:27:10.559204 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3afbab4e_d827_41c5_b3a7_65061daf2eeb.slice/crio-f2c55a0a1288ad768f7aad35272269358d20d9bd7027d61a4a1546dc9524b1da WatchSource:0}: Error finding container f2c55a0a1288ad768f7aad35272269358d20d9bd7027d61a4a1546dc9524b1da: Status 404 returned error can't find the container with id f2c55a0a1288ad768f7aad35272269358d20d9bd7027d61a4a1546dc9524b1da Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.596537 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp"] Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.614609 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:10 crc kubenswrapper[4744]: E1003 16:27:10.615056 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:11.115037606 +0000 UTC m=+157.394913492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:10 crc kubenswrapper[4744]: W1003 16:27:10.692565 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4cdd3c3_ae45_4dad_a55e_13dea19f1f5a.slice/crio-326c049f3a124f05ad8f194e6c8bf194dd249750bf7cac620bfa03a036916e9e WatchSource:0}: Error finding container 326c049f3a124f05ad8f194e6c8bf194dd249750bf7cac620bfa03a036916e9e: Status 404 returned error can't find the container with id 326c049f3a124f05ad8f194e6c8bf194dd249750bf7cac620bfa03a036916e9e Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.715474 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:10 crc kubenswrapper[4744]: E1003 16:27:10.715849 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:11.215832473 +0000 UTC m=+157.495708369 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.804355 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.804421 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.813058 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.817660 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:10 crc kubenswrapper[4744]: E1003 16:27:10.818017 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:11.318004498 +0000 UTC m=+157.597880394 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:10 crc kubenswrapper[4744]: I1003 16:27:10.919222 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:10 crc kubenswrapper[4744]: E1003 16:27:10.921297 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:11.421275963 +0000 UTC m=+157.701151859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.022005 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:11 crc kubenswrapper[4744]: E1003 16:27:11.022917 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:11.522896922 +0000 UTC m=+157.802772818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.124455 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:11 crc kubenswrapper[4744]: E1003 16:27:11.124720 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:11.624673526 +0000 UTC m=+157.904549422 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.125102 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:11 crc kubenswrapper[4744]: E1003 16:27:11.125598 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:11.625585071 +0000 UTC m=+157.905460967 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.226845 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:11 crc kubenswrapper[4744]: E1003 16:27:11.227674 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:11.727641843 +0000 UTC m=+158.007517739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.254345 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" event={"ID":"36b56a51-ac0b-437b-8f4f-e42476ed6ddd","Type":"ContainerStarted","Data":"12f8f094d36198b80c89e2bdb9a9d573db037df8d8736c6e6967aa06291db570"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.264835 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb" event={"ID":"00092b0a-d430-4d0b-a90d-3a5eb74cbb88","Type":"ContainerStarted","Data":"7c845066e921e912025931dfa54b5b2570a6a0c31e41c2b7722ac12e2cf91616"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.264887 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb" event={"ID":"00092b0a-d430-4d0b-a90d-3a5eb74cbb88","Type":"ContainerStarted","Data":"db9165efacb55f8061214ab7349927d295b014de051a2664eeafd770f16daf17"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.272861 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:11 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:11 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:11 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.272945 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.277938 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" event={"ID":"4e224494-c68b-4f4a-a32c-b631f35ad663","Type":"ContainerStarted","Data":"888071492fe2aa00fee35206f8f6443b4a84be9a6bb57906e5717d1b0479449c"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.284373 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp" event={"ID":"b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a","Type":"ContainerStarted","Data":"326c049f3a124f05ad8f194e6c8bf194dd249750bf7cac620bfa03a036916e9e"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.300408 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cffnb" podStartSLOduration=136.300378473 podStartE2EDuration="2m16.300378473s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:11.293892715 +0000 UTC m=+157.573768611" watchObservedRunningTime="2025-10-03 16:27:11.300378473 +0000 UTC m=+157.580254369" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.320645 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx" event={"ID":"55164953-87bb-4683-8afd-496bbbf6385a","Type":"ContainerStarted","Data":"072c6f856bbb838035310d450c7bcca52884ab23a86ff77b52b1322de9884be7"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.333234 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:11 crc kubenswrapper[4744]: E1003 16:27:11.333671 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:11.833653843 +0000 UTC m=+158.113529739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.383296 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" event={"ID":"4f39c88b-8d2a-4a19-921e-14d8d172ad20","Type":"ContainerStarted","Data":"9daf5dec3e1dfc007bd7c8e2369b05cb67283ba709124f784ec8c34066e25897"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.384881 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.392646 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" event={"ID":"e93027df-377c-47a1-b2c6-9cd42db258ab","Type":"ContainerStarted","Data":"1dfeb14f6c82ce40a77e9220bbd992d36fe7b5a47a0f3de8b8e953616ae712d7"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.392688 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" event={"ID":"e93027df-377c-47a1-b2c6-9cd42db258ab","Type":"ContainerStarted","Data":"d882b061c427e895efae12bc8da92a91f023aab4ad4dd5ac84ed643d6c75695a"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.394074 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.407679 4744 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-6jxqm container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" start-of-body= Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.407748 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" podUID="4f39c88b-8d2a-4a19-921e-14d8d172ad20" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.412312 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh" event={"ID":"37e50c68-0215-4d7e-aea2-e7e85dcd9b4b","Type":"ContainerStarted","Data":"65e6cb54da7c5df4181eac28570a2ff7ecfca34e13d8089973f0f3956df9f93b"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.412360 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh" event={"ID":"37e50c68-0215-4d7e-aea2-e7e85dcd9b4b","Type":"ContainerStarted","Data":"950fa239571988c398e15e8581b14debbb0411c30e1d474150e483bca595bd84"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.422641 4744 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-q6hdh container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.422688 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" podUID="e93027df-377c-47a1-b2c6-9cd42db258ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.428078 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w4wdx" podStartSLOduration=135.428042395 podStartE2EDuration="2m15.428042395s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:11.358103712 +0000 UTC m=+157.637979608" watchObservedRunningTime="2025-10-03 16:27:11.428042395 +0000 UTC m=+157.707918291" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.435176 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:11 crc kubenswrapper[4744]: E1003 16:27:11.435408 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:11.935371845 +0000 UTC m=+158.215247741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.435855 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:11 crc kubenswrapper[4744]: E1003 16:27:11.437704 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:11.937690179 +0000 UTC m=+158.217566075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.475571 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" podStartSLOduration=136.475549334 podStartE2EDuration="2m16.475549334s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:11.431060287 +0000 UTC m=+157.710936183" watchObservedRunningTime="2025-10-03 16:27:11.475549334 +0000 UTC m=+157.755425230" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.516684 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" event={"ID":"62497d3a-2816-455c-824e-d908c11e71e1","Type":"ContainerStarted","Data":"0db7190931f983ae9e46224d23780971bab6bbee1025822a6c21787830fb70a4"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.528592 4744 generic.go:334] "Generic (PLEG): container finished" podID="37928f66-a94f-4241-b984-e0729ee6dbf0" containerID="6abe209af7d549a12c927a72ca152a0082ee4f24bdaf5bd94c8e16b6dd3627e0" exitCode=0 Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.528687 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" event={"ID":"37928f66-a94f-4241-b984-e0729ee6dbf0","Type":"ContainerDied","Data":"6abe209af7d549a12c927a72ca152a0082ee4f24bdaf5bd94c8e16b6dd3627e0"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.553001 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" podStartSLOduration=135.552980383 podStartE2EDuration="2m15.552980383s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:11.471594516 +0000 UTC m=+157.751470412" watchObservedRunningTime="2025-10-03 16:27:11.552980383 +0000 UTC m=+157.832856309" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.555026 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:11 crc kubenswrapper[4744]: E1003 16:27:11.555295 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:12.055273955 +0000 UTC m=+158.335149851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.556257 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:11 crc kubenswrapper[4744]: E1003 16:27:11.559256 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:12.059238254 +0000 UTC m=+158.339114330 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.566103 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-dmld4" event={"ID":"3668ee6f-1314-4aec-ac55-9b55998f3c75","Type":"ContainerStarted","Data":"276ba9525328fe6134cd2daa6421db4d28fc6f29a6af720dade4045efb9a5573"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.567654 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" podStartSLOduration=136.567623933 podStartE2EDuration="2m16.567623933s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:11.552616883 +0000 UTC m=+157.832492779" watchObservedRunningTime="2025-10-03 16:27:11.567623933 +0000 UTC m=+157.847499829" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.579437 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-v5snt" event={"ID":"d61086b0-b78f-4239-8d9a-49aba28c5820","Type":"ContainerStarted","Data":"e5a2e9f9e8e08257f04a519e7f0d4d938d3a17733f7b72d5b92b5764268151e4"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.611686 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" event={"ID":"a96e3d62-e019-4387-9ee1-4ea8dd90189c","Type":"ContainerStarted","Data":"f65074654cb57b64fa3dc1d5e691caf6881243540cefda9251aa29605d0f3aa7"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.612958 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.643071 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" event={"ID":"8535c47b-978d-43cf-a561-50f78e3f927d","Type":"ContainerStarted","Data":"0aa636402c695be845c99aef54de30352e244c41004299518f156879bbe50a4c"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.643131 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" event={"ID":"8535c47b-978d-43cf-a561-50f78e3f927d","Type":"ContainerStarted","Data":"fb1e0761e287414eaafc93330036de0015159ba0bace4fc580f6f50049ffb661"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.644199 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.647340 4744 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wzkjd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.647414 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" podUID="8535c47b-978d-43cf-a561-50f78e3f927d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.658839 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.659231 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-v5snt" podStartSLOduration=136.659202108 podStartE2EDuration="2m16.659202108s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:11.658352655 +0000 UTC m=+157.938228551" watchObservedRunningTime="2025-10-03 16:27:11.659202108 +0000 UTC m=+157.939078004" Oct 03 16:27:11 crc kubenswrapper[4744]: E1003 16:27:11.662951 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:12.16292573 +0000 UTC m=+158.442801626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.660161 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-dmld4" podStartSLOduration=7.660151884 podStartE2EDuration="7.660151884s" podCreationTimestamp="2025-10-03 16:27:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:11.621460816 +0000 UTC m=+157.901336712" watchObservedRunningTime="2025-10-03 16:27:11.660151884 +0000 UTC m=+157.940027780" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.673956 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.702000 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" podStartSLOduration=135.701946967 podStartE2EDuration="2m15.701946967s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:11.698105592 +0000 UTC m=+157.977981488" watchObservedRunningTime="2025-10-03 16:27:11.701946967 +0000 UTC m=+157.981822863" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.715460 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dpx8q" event={"ID":"59f447fc-64dd-4628-9a42-a821c990643a","Type":"ContainerStarted","Data":"77092bc93e0f54d46bfbed298849b6fedaf696661552a59d16234864c6883adf"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.738022 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" event={"ID":"a0c1475e-1086-40fc-800e-91f5156c92f9","Type":"ContainerStarted","Data":"4398013795b433df3a5222453f123b4bff3bdffa6b560128017dd391c68e6b59"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.761944 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:11 crc kubenswrapper[4744]: E1003 16:27:11.763168 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:12.263148762 +0000 UTC m=+158.543024658 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.778616 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cj6h4" podStartSLOduration=135.778596574 podStartE2EDuration="2m15.778596574s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:11.77516061 +0000 UTC m=+158.055036506" watchObservedRunningTime="2025-10-03 16:27:11.778596574 +0000 UTC m=+158.058472470" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.813964 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qmddz" event={"ID":"3afbab4e-d827-41c5-b3a7-65061daf2eeb","Type":"ContainerStarted","Data":"f2c55a0a1288ad768f7aad35272269358d20d9bd7027d61a4a1546dc9524b1da"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.864958 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:11 crc kubenswrapper[4744]: E1003 16:27:11.866745 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:12.366720795 +0000 UTC m=+158.646596691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.866905 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg" event={"ID":"dad9df42-c989-4fdf-8f19-ed7873979522","Type":"ContainerStarted","Data":"8566959721e668925767adf20d1b4636ce2d1f21a6fed61a27d2c682a03ab8eb"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.866962 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg" event={"ID":"dad9df42-c989-4fdf-8f19-ed7873979522","Type":"ContainerStarted","Data":"b1b0ee696ae86bf273c004483da9ca9a534739b5c1857c7d389f820bb3000be3"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.886619 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dfzqw" event={"ID":"fe526e45-7a63-4289-a861-2c10510a0268","Type":"ContainerStarted","Data":"184786f2f1fbaf274b3e20158ff444ba9c8a15e7b52fe5153d3e88a411bb8dc6"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.886693 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dfzqw" event={"ID":"fe526e45-7a63-4289-a861-2c10510a0268","Type":"ContainerStarted","Data":"2c4afe3df424570307a41bd157463b8cf0cf347a824eb338b7e492d336adc27c"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.890898 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-dpx8q" podStartSLOduration=135.890885556 podStartE2EDuration="2m15.890885556s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:11.879609007 +0000 UTC m=+158.159484903" watchObservedRunningTime="2025-10-03 16:27:11.890885556 +0000 UTC m=+158.170761452" Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.938070 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" event={"ID":"c3df59a0-c796-4915-b17b-ecfd5f0d40b1","Type":"ContainerStarted","Data":"08cba4933d0e11d9569679d5c5fdcdf508bedf24728a48783ee81b17c6acaf65"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.938140 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" event={"ID":"c3df59a0-c796-4915-b17b-ecfd5f0d40b1","Type":"ContainerStarted","Data":"7fd6d15bea47d5798df18d427af1c505bf189c5f1e16df87daf803bff1e0b436"} Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.966633 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:11 crc kubenswrapper[4744]: E1003 16:27:11.967070 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:12.467049949 +0000 UTC m=+158.746925845 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:11 crc kubenswrapper[4744]: I1003 16:27:11.988579 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb" event={"ID":"cbbb76ba-cb04-4559-a927-00e3ad2bcea0","Type":"ContainerStarted","Data":"d7f6f47d5315ea03d9dd7a53fe2a7be8c77b1fdd913abb6c382080e2ef96b9a9"} Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.014333 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmzm9" event={"ID":"c41f7473-52d9-4d32-9364-ed61efb65af4","Type":"ContainerStarted","Data":"bb41e0e8749f9fea86a2252769efa27145face082f839285c558f6de4516e72d"} Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.017011 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2kpgr" podStartSLOduration=137.016997666 podStartE2EDuration="2m17.016997666s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:11.938222841 +0000 UTC m=+158.218098737" watchObservedRunningTime="2025-10-03 16:27:12.016997666 +0000 UTC m=+158.296873562" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.017450 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wc4mg" podStartSLOduration=136.017443308 podStartE2EDuration="2m16.017443308s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:12.014760704 +0000 UTC m=+158.294636600" watchObservedRunningTime="2025-10-03 16:27:12.017443308 +0000 UTC m=+158.297319204" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.024856 4744 generic.go:334] "Generic (PLEG): container finished" podID="5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7" containerID="24524adcd0257b8674d4a17b8091c3bef1ad5c0af89f34f652cdf6c86dc71a6b" exitCode=0 Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.024934 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" event={"ID":"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7","Type":"ContainerDied","Data":"24524adcd0257b8674d4a17b8091c3bef1ad5c0af89f34f652cdf6c86dc71a6b"} Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.044613 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94qmg" event={"ID":"71471281-c2f7-42fb-8cde-6d18b8dba543","Type":"ContainerStarted","Data":"8bb24f93c4fd71e05bb068c6f527593eaf007a2bf8ffee70c92360f219982220"} Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.044683 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94qmg" event={"ID":"71471281-c2f7-42fb-8cde-6d18b8dba543","Type":"ContainerStarted","Data":"fc972e16d22994e1379d4a801a4677c81a80cb6adbeb0e1d94f8f1592877b14a"} Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.052317 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" event={"ID":"f4865f7c-50e4-4159-8c57-42c69bd4efee","Type":"ContainerStarted","Data":"d63e5fb93906f3a15b3fa5772ec070321dcdab11c4205e28c41df5e4e6018cd0"} Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.072815 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:12 crc kubenswrapper[4744]: E1003 16:27:12.073201 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:12.573168112 +0000 UTC m=+158.853044008 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.073361 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:12 crc kubenswrapper[4744]: E1003 16:27:12.075028 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:12.575006162 +0000 UTC m=+158.854882048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.084282 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-nbsh6" event={"ID":"16bea17d-98bf-4807-b4b3-3802d3c83e21","Type":"ContainerStarted","Data":"879e17c57f902f841653cb881e1edcacd8bc4a0130ba6b8960a503d06e20e9bb"} Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.114362 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v" event={"ID":"a5c7a90d-1be3-47c3-a945-e7b872888de4","Type":"ContainerStarted","Data":"8c8ee44888b1bfbe04441b5a2c266e3e924ef013f315354652e4712eee021b3a"} Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.115652 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ltmsn" podStartSLOduration=136.115632274 podStartE2EDuration="2m16.115632274s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:12.114275967 +0000 UTC m=+158.394151863" watchObservedRunningTime="2025-10-03 16:27:12.115632274 +0000 UTC m=+158.395508160" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.125717 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-s96bg" event={"ID":"0784e651-ae4d-417f-9353-5550c790b5a4","Type":"ContainerStarted","Data":"26241a0458415e1e3e2fdceaa81fe787a9618c2cf53585cd7a4f212ccfb22299"} Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.145872 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" event={"ID":"cfcd176c-f3fd-419f-92ad-f7fdbcab6270","Type":"ContainerStarted","Data":"41ae90d5fc75256882279e151f4940c81ba81084375f4ba5a2c808f95f4a4a89"} Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.145935 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.145950 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" event={"ID":"cfcd176c-f3fd-419f-92ad-f7fdbcab6270","Type":"ContainerStarted","Data":"286a60c0444c4cf67a80c02e6137b130e8c68a28048b05284a094712b3b7cfe1"} Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.162251 4744 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9hm6h container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.162333 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" podUID="cfcd176c-f3fd-419f-92ad-f7fdbcab6270" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.167186 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7" event={"ID":"a0b52a53-304a-4664-b8d7-9434f969dac3","Type":"ContainerStarted","Data":"9d7fca680199edf72aa3c3a7064b236513cfc91865eaf9cd10abf8864114419f"} Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.172331 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-ksbgj" podStartSLOduration=137.172308584 podStartE2EDuration="2m17.172308584s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:12.17142121 +0000 UTC m=+158.451297096" watchObservedRunningTime="2025-10-03 16:27:12.172308584 +0000 UTC m=+158.452184480" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.175289 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:12 crc kubenswrapper[4744]: E1003 16:27:12.176837 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:12.676815557 +0000 UTC m=+158.956691453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.201845 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6t5pf" event={"ID":"b2a7d1a9-0abe-4150-8c23-c9f592467ba5","Type":"ContainerStarted","Data":"a6982db73096507694cda89066ac2de9b4e1cbbb523b5c614c7a57757b5aa03f"} Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.214479 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-k2tzp" event={"ID":"3c105e35-cb55-4769-a2e3-5e22ed8478c3","Type":"ContainerStarted","Data":"9e1ee99ae569bbedca70722d3d920822a72f994170da37e29183ba3cdbee2e82"} Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.214545 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.222456 4744 patch_prober.go:28] interesting pod/console-operator-58897d9998-k2tzp container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.222541 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-k2tzp" podUID="3c105e35-cb55-4769-a2e3-5e22ed8478c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.222668 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-thbg6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.222711 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-thbg6" podUID="6ac85db3-283c-43f9-9925-b8a0f8354181" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.241329 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vrs7" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.268655 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:12 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:12 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:12 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.268994 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.281845 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb" podStartSLOduration=136.28182765 podStartE2EDuration="2m16.28182765s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:12.227429402 +0000 UTC m=+158.507305308" watchObservedRunningTime="2025-10-03 16:27:12.28182765 +0000 UTC m=+158.561703546" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.282626 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:12 crc kubenswrapper[4744]: E1003 16:27:12.321747 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:12.821705851 +0000 UTC m=+159.101581747 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.370279 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-94qmg" podStartSLOduration=137.370244579 podStartE2EDuration="2m17.370244579s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:12.34725623 +0000 UTC m=+158.627132126" watchObservedRunningTime="2025-10-03 16:27:12.370244579 +0000 UTC m=+158.650120475" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.390254 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:12 crc kubenswrapper[4744]: E1003 16:27:12.390810 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:12.89078386 +0000 UTC m=+159.170659756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.397907 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qmzm9" podStartSLOduration=136.397882015 podStartE2EDuration="2m16.397882015s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:12.388989991 +0000 UTC m=+158.668865887" watchObservedRunningTime="2025-10-03 16:27:12.397882015 +0000 UTC m=+158.677757911" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.437291 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-k2tzp" podStartSLOduration=137.437266932 podStartE2EDuration="2m17.437266932s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:12.436025668 +0000 UTC m=+158.715901584" watchObservedRunningTime="2025-10-03 16:27:12.437266932 +0000 UTC m=+158.717142828" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.495671 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:12 crc kubenswrapper[4744]: E1003 16:27:12.496120 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:12.996106382 +0000 UTC m=+159.275982268 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.559638 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v" podStartSLOduration=136.559619219 podStartE2EDuration="2m16.559619219s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:12.559239198 +0000 UTC m=+158.839115094" watchObservedRunningTime="2025-10-03 16:27:12.559619219 +0000 UTC m=+158.839495115" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.581107 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-s96bg" podStartSLOduration=8.581071936 podStartE2EDuration="8.581071936s" podCreationTimestamp="2025-10-03 16:27:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:12.580751817 +0000 UTC m=+158.860627713" watchObservedRunningTime="2025-10-03 16:27:12.581071936 +0000 UTC m=+158.860947832" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.597738 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:12 crc kubenswrapper[4744]: E1003 16:27:12.598630 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:13.098608835 +0000 UTC m=+159.378484731 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.626624 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6t5pf" podStartSLOduration=136.626608311 podStartE2EDuration="2m16.626608311s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:12.625965754 +0000 UTC m=+158.905841650" watchObservedRunningTime="2025-10-03 16:27:12.626608311 +0000 UTC m=+158.906484207" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.699803 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:12 crc kubenswrapper[4744]: E1003 16:27:12.700154 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:13.200141683 +0000 UTC m=+159.480017579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.716484 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w8sw7" podStartSLOduration=136.716456329 podStartE2EDuration="2m16.716456329s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:12.714104075 +0000 UTC m=+158.993979971" watchObservedRunningTime="2025-10-03 16:27:12.716456329 +0000 UTC m=+158.996332225" Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.800726 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:12 crc kubenswrapper[4744]: E1003 16:27:12.801146 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:13.301121645 +0000 UTC m=+159.580997541 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:12 crc kubenswrapper[4744]: I1003 16:27:12.902304 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:12 crc kubenswrapper[4744]: E1003 16:27:12.902884 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:13.402863357 +0000 UTC m=+159.682739413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.003590 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:13 crc kubenswrapper[4744]: E1003 16:27:13.003819 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:13.503785178 +0000 UTC m=+159.783661084 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.004039 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:13 crc kubenswrapper[4744]: E1003 16:27:13.004432 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:13.504424115 +0000 UTC m=+159.784300011 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.104727 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:13 crc kubenswrapper[4744]: E1003 16:27:13.105069 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:13.605031428 +0000 UTC m=+159.884907324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.105671 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:13 crc kubenswrapper[4744]: E1003 16:27:13.106144 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:13.606129088 +0000 UTC m=+159.886004984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.206336 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:13 crc kubenswrapper[4744]: E1003 16:27:13.206451 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:13.706435411 +0000 UTC m=+159.986311307 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.206814 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:13 crc kubenswrapper[4744]: E1003 16:27:13.207143 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:13.707135541 +0000 UTC m=+159.987011437 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.220895 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-nbsh6" event={"ID":"16bea17d-98bf-4807-b4b3-3802d3c83e21","Type":"ContainerStarted","Data":"b7b98654ad0e74d81adc7120b4a0b91e0fc37ff9ea69be500114015191615f81"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.222836 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp" event={"ID":"b4cdd3c3-ae45-4dad-a55e-13dea19f1f5a","Type":"ContainerStarted","Data":"0aaa8a4f9eb5fb060badd4e0d9845f8f38ae235e9c2d9cddfb1f02479375ca7c"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.225038 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dpx8q" event={"ID":"59f447fc-64dd-4628-9a42-a821c990643a","Type":"ContainerStarted","Data":"7bbbf2070e87485e6e52cb8f9b0ec79dfee34355f35053b979444326f9757dbc"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.227383 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" event={"ID":"4e224494-c68b-4f4a-a32c-b631f35ad663","Type":"ContainerStarted","Data":"31aa0dfa0a68d2d4d2f5c405a5e6906efa1db7a96241014a59e4767e7c331168"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.229073 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-s96bg" event={"ID":"0784e651-ae4d-417f-9353-5550c790b5a4","Type":"ContainerStarted","Data":"1b06730786c7efddf7fbc6e952540f049e119cf5d047a7f3ac0d295856d62e8e"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.231788 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" event={"ID":"37928f66-a94f-4241-b984-e0729ee6dbf0","Type":"ContainerStarted","Data":"59f9016f8540864eb83d79dc0e093f111bcfa1d083a47b8fc4e416b82bbd9b10"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.232169 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.234005 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" event={"ID":"36b56a51-ac0b-437b-8f4f-e42476ed6ddd","Type":"ContainerStarted","Data":"4d862c88a2db4af5aa173bb3bc56272c9bd494a34cabebc0931ad44d8744d8aa"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.236298 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5js5v" event={"ID":"a5c7a90d-1be3-47c3-a945-e7b872888de4","Type":"ContainerStarted","Data":"63fc374170cdc1f750bc591f0797809c960b70248d9258a29b4e2f6808ff88d5"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.238572 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-k2tzp" event={"ID":"3c105e35-cb55-4769-a2e3-5e22ed8478c3","Type":"ContainerStarted","Data":"051ad6f686cffee655e8b9eb1b13a559f3a7c86a26ddb1798b0b776a512bb48c"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.239600 4744 patch_prober.go:28] interesting pod/console-operator-58897d9998-k2tzp container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.239699 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-k2tzp" podUID="3c105e35-cb55-4769-a2e3-5e22ed8478c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.241254 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" event={"ID":"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7","Type":"ContainerStarted","Data":"39a24648558600e854821b72b5055f99e8eeecb114540a4746d43414f1da89b1"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.241283 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" event={"ID":"5cfe49a4-a424-4e8c-82f6-9a38aabaa5c7","Type":"ContainerStarted","Data":"3bea0717fb4f0e5fef971a28a7dc93cd9b63bda4547322557a05c41a808bd294"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.243115 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qmddz" event={"ID":"3afbab4e-d827-41c5-b3a7-65061daf2eeb","Type":"ContainerStarted","Data":"1601ec5b6346a7dc85d2b0a4c8f7312f02ba9ba505cb1df24e2aceb37000042c"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.243182 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qmddz" event={"ID":"3afbab4e-d827-41c5-b3a7-65061daf2eeb","Type":"ContainerStarted","Data":"13c22369cdaedc24def0fbd81dba8088b492182fe5335607ef8b88135b3e7af3"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.243206 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-qmddz" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.245811 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh" event={"ID":"37e50c68-0215-4d7e-aea2-e7e85dcd9b4b","Type":"ContainerStarted","Data":"39a0f9a2b47383d1929193fd86befd302c2b70e5e64f009491e890172c93a5a5"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.246010 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.247804 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-dfzqw" event={"ID":"fe526e45-7a63-4289-a861-2c10510a0268","Type":"ContainerStarted","Data":"f9dff64cef2460e26b31ff865a2252cfd82572a655cc380081f45ba222902531"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.250122 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mhrlb" event={"ID":"cbbb76ba-cb04-4559-a927-00e3ad2bcea0","Type":"ContainerStarted","Data":"3d8dde2cf3c4c338a76d3efdfeb0e4d079410347873d4a40ae26bc9cda83f8f7"} Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.252408 4744 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9hm6h container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.252475 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" podUID="cfcd176c-f3fd-419f-92ad-f7fdbcab6270" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.263097 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-q6hdh" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.264640 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" podStartSLOduration=137.264614163 podStartE2EDuration="2m17.264614163s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:12.78776776 +0000 UTC m=+159.067643656" watchObservedRunningTime="2025-10-03 16:27:13.264614163 +0000 UTC m=+159.544490059" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.266244 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-nbsh6" podStartSLOduration=137.266230317 podStartE2EDuration="2m17.266230317s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:13.262336731 +0000 UTC m=+159.542212627" watchObservedRunningTime="2025-10-03 16:27:13.266230317 +0000 UTC m=+159.546106213" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.267618 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:13 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:13 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:13 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.267674 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.271337 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.308361 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:13 crc kubenswrapper[4744]: E1003 16:27:13.308631 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:13.808593466 +0000 UTC m=+160.088469362 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.309893 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:13 crc kubenswrapper[4744]: E1003 16:27:13.312296 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:13.812278577 +0000 UTC m=+160.092154473 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.342342 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-qmddz" podStartSLOduration=9.342312248 podStartE2EDuration="9.342312248s" podCreationTimestamp="2025-10-03 16:27:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:13.326188267 +0000 UTC m=+159.606064173" watchObservedRunningTime="2025-10-03 16:27:13.342312248 +0000 UTC m=+159.622188144" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.403014 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" podStartSLOduration=138.402991678 podStartE2EDuration="2m18.402991678s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:13.40195369 +0000 UTC m=+159.681829596" watchObservedRunningTime="2025-10-03 16:27:13.402991678 +0000 UTC m=+159.682867584" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.419262 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:13 crc kubenswrapper[4744]: E1003 16:27:13.419713 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:13.919695035 +0000 UTC m=+160.199570931 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.459214 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" podStartSLOduration=138.459193456 podStartE2EDuration="2m18.459193456s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:13.438036727 +0000 UTC m=+159.717912623" watchObservedRunningTime="2025-10-03 16:27:13.459193456 +0000 UTC m=+159.739069352" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.487259 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh" podStartSLOduration=137.487237193 podStartE2EDuration="2m17.487237193s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:13.478022931 +0000 UTC m=+159.757898827" watchObservedRunningTime="2025-10-03 16:27:13.487237193 +0000 UTC m=+159.767113089" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.522348 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:13 crc kubenswrapper[4744]: E1003 16:27:13.523138 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:14.023121554 +0000 UTC m=+160.302997450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.539678 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rspbp" podStartSLOduration=137.539654086 podStartE2EDuration="2m17.539654086s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:13.538053803 +0000 UTC m=+159.817929699" watchObservedRunningTime="2025-10-03 16:27:13.539654086 +0000 UTC m=+159.819529983" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.626176 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:13 crc kubenswrapper[4744]: E1003 16:27:13.626764 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:14.126738709 +0000 UTC m=+160.406614605 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.660555 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cgkrq" podStartSLOduration=137.660531313 podStartE2EDuration="2m17.660531313s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:13.600379068 +0000 UTC m=+159.880254964" watchObservedRunningTime="2025-10-03 16:27:13.660531313 +0000 UTC m=+159.940407209" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.726033 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-dfzqw" podStartSLOduration=138.726008284 podStartE2EDuration="2m18.726008284s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:13.663147265 +0000 UTC m=+159.943023161" watchObservedRunningTime="2025-10-03 16:27:13.726008284 +0000 UTC m=+160.005884180" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.732432 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:13 crc kubenswrapper[4744]: E1003 16:27:13.732811 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:14.23279544 +0000 UTC m=+160.512671336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.838136 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:13 crc kubenswrapper[4744]: E1003 16:27:13.838231 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:14.338185523 +0000 UTC m=+160.618061419 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.839011 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:13 crc kubenswrapper[4744]: E1003 16:27:13.839466 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:14.339457778 +0000 UTC m=+160.619333674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.919235 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7vn7c"] Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.920388 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.923822 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.940855 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:13 crc kubenswrapper[4744]: E1003 16:27:13.941214 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:14.44118497 +0000 UTC m=+160.721060866 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:13 crc kubenswrapper[4744]: I1003 16:27:13.962964 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7vn7c"] Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.042908 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.043099 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7570f2af-fefc-4726-a9ec-1c27d7cefe43-utilities\") pod \"community-operators-7vn7c\" (UID: \"7570f2af-fefc-4726-a9ec-1c27d7cefe43\") " pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.043203 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7570f2af-fefc-4726-a9ec-1c27d7cefe43-catalog-content\") pod \"community-operators-7vn7c\" (UID: \"7570f2af-fefc-4726-a9ec-1c27d7cefe43\") " pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.043302 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpl6l\" (UniqueName: \"kubernetes.io/projected/7570f2af-fefc-4726-a9ec-1c27d7cefe43-kube-api-access-kpl6l\") pod \"community-operators-7vn7c\" (UID: \"7570f2af-fefc-4726-a9ec-1c27d7cefe43\") " pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:27:14 crc kubenswrapper[4744]: E1003 16:27:14.043690 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:14.543676704 +0000 UTC m=+160.823552600 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.118965 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nh227"] Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.120053 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.121635 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.141571 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nh227"] Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.145426 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.145690 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7570f2af-fefc-4726-a9ec-1c27d7cefe43-utilities\") pod \"community-operators-7vn7c\" (UID: \"7570f2af-fefc-4726-a9ec-1c27d7cefe43\") " pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.145727 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7570f2af-fefc-4726-a9ec-1c27d7cefe43-catalog-content\") pod \"community-operators-7vn7c\" (UID: \"7570f2af-fefc-4726-a9ec-1c27d7cefe43\") " pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.145755 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpl6l\" (UniqueName: \"kubernetes.io/projected/7570f2af-fefc-4726-a9ec-1c27d7cefe43-kube-api-access-kpl6l\") pod \"community-operators-7vn7c\" (UID: \"7570f2af-fefc-4726-a9ec-1c27d7cefe43\") " pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:27:14 crc kubenswrapper[4744]: E1003 16:27:14.145860 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:14.645816988 +0000 UTC m=+160.925692884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.146272 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7570f2af-fefc-4726-a9ec-1c27d7cefe43-utilities\") pod \"community-operators-7vn7c\" (UID: \"7570f2af-fefc-4726-a9ec-1c27d7cefe43\") " pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.146330 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7570f2af-fefc-4726-a9ec-1c27d7cefe43-catalog-content\") pod \"community-operators-7vn7c\" (UID: \"7570f2af-fefc-4726-a9ec-1c27d7cefe43\") " pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.192540 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpl6l\" (UniqueName: \"kubernetes.io/projected/7570f2af-fefc-4726-a9ec-1c27d7cefe43-kube-api-access-kpl6l\") pod \"community-operators-7vn7c\" (UID: \"7570f2af-fefc-4726-a9ec-1c27d7cefe43\") " pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.233650 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wzkjd" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.234131 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.248108 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-catalog-content\") pod \"certified-operators-nh227\" (UID: \"18d30e6e-1a94-4d75-8b09-e01c3eb2575e\") " pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.248165 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-utilities\") pod \"certified-operators-nh227\" (UID: \"18d30e6e-1a94-4d75-8b09-e01c3eb2575e\") " pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.248210 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk9qh\" (UniqueName: \"kubernetes.io/projected/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-kube-api-access-hk9qh\") pod \"certified-operators-nh227\" (UID: \"18d30e6e-1a94-4d75-8b09-e01c3eb2575e\") " pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.248253 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:14 crc kubenswrapper[4744]: E1003 16:27:14.248624 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:14.74860762 +0000 UTC m=+161.028483506 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.279421 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:14 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:14 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:14 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.279511 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.350107 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:14 crc kubenswrapper[4744]: E1003 16:27:14.350416 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:14.850396884 +0000 UTC m=+161.130272780 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.350513 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-catalog-content\") pod \"certified-operators-nh227\" (UID: \"18d30e6e-1a94-4d75-8b09-e01c3eb2575e\") " pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.350547 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-utilities\") pod \"certified-operators-nh227\" (UID: \"18d30e6e-1a94-4d75-8b09-e01c3eb2575e\") " pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.350593 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk9qh\" (UniqueName: \"kubernetes.io/projected/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-kube-api-access-hk9qh\") pod \"certified-operators-nh227\" (UID: \"18d30e6e-1a94-4d75-8b09-e01c3eb2575e\") " pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.350622 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:14 crc kubenswrapper[4744]: E1003 16:27:14.350964 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:14.85095538 +0000 UTC m=+161.130831276 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.352543 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" event={"ID":"36b56a51-ac0b-437b-8f4f-e42476ed6ddd","Type":"ContainerStarted","Data":"474ad978c4167dfb6c27661674de1debf05683a0b5a259f5f9f3736839ce5eeb"} Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.358732 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-catalog-content\") pod \"certified-operators-nh227\" (UID: \"18d30e6e-1a94-4d75-8b09-e01c3eb2575e\") " pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.359428 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-utilities\") pod \"certified-operators-nh227\" (UID: \"18d30e6e-1a94-4d75-8b09-e01c3eb2575e\") " pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.359847 4744 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9hm6h container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.359885 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" podUID="cfcd176c-f3fd-419f-92ad-f7fdbcab6270" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.362280 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mv2tt"] Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.363963 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.382606 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mv2tt"] Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.446012 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-k2tzp" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.455423 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.456350 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxm8k\" (UniqueName: \"kubernetes.io/projected/8be1fdf2-9a41-455e-8315-459cca9f8a09-kube-api-access-lxm8k\") pod \"community-operators-mv2tt\" (UID: \"8be1fdf2-9a41-455e-8315-459cca9f8a09\") " pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:27:14 crc kubenswrapper[4744]: E1003 16:27:14.456995 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:14.956967009 +0000 UTC m=+161.236842905 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.457391 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be1fdf2-9a41-455e-8315-459cca9f8a09-utilities\") pod \"community-operators-mv2tt\" (UID: \"8be1fdf2-9a41-455e-8315-459cca9f8a09\") " pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.457639 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.457945 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be1fdf2-9a41-455e-8315-459cca9f8a09-catalog-content\") pod \"community-operators-mv2tt\" (UID: \"8be1fdf2-9a41-455e-8315-459cca9f8a09\") " pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:27:14 crc kubenswrapper[4744]: E1003 16:27:14.461541 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:14.961526224 +0000 UTC m=+161.241402120 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.469722 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk9qh\" (UniqueName: \"kubernetes.io/projected/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-kube-api-access-hk9qh\") pod \"certified-operators-nh227\" (UID: \"18d30e6e-1a94-4d75-8b09-e01c3eb2575e\") " pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.546646 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v5tr5"] Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.556064 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.559133 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v5tr5"] Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.560078 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.560463 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be1fdf2-9a41-455e-8315-459cca9f8a09-catalog-content\") pod \"community-operators-mv2tt\" (UID: \"8be1fdf2-9a41-455e-8315-459cca9f8a09\") " pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.560639 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxm8k\" (UniqueName: \"kubernetes.io/projected/8be1fdf2-9a41-455e-8315-459cca9f8a09-kube-api-access-lxm8k\") pod \"community-operators-mv2tt\" (UID: \"8be1fdf2-9a41-455e-8315-459cca9f8a09\") " pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:27:14 crc kubenswrapper[4744]: E1003 16:27:14.560768 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:15.060737168 +0000 UTC m=+161.340613064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.560824 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be1fdf2-9a41-455e-8315-459cca9f8a09-utilities\") pod \"community-operators-mv2tt\" (UID: \"8be1fdf2-9a41-455e-8315-459cca9f8a09\") " pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.561376 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be1fdf2-9a41-455e-8315-459cca9f8a09-utilities\") pod \"community-operators-mv2tt\" (UID: \"8be1fdf2-9a41-455e-8315-459cca9f8a09\") " pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.564189 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be1fdf2-9a41-455e-8315-459cca9f8a09-catalog-content\") pod \"community-operators-mv2tt\" (UID: \"8be1fdf2-9a41-455e-8315-459cca9f8a09\") " pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.608923 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxm8k\" (UniqueName: \"kubernetes.io/projected/8be1fdf2-9a41-455e-8315-459cca9f8a09-kube-api-access-lxm8k\") pod \"community-operators-mv2tt\" (UID: \"8be1fdf2-9a41-455e-8315-459cca9f8a09\") " pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.662189 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vqzf\" (UniqueName: \"kubernetes.io/projected/f8087c3f-5904-4338-8dcf-779cf0dfa81c-kube-api-access-7vqzf\") pod \"certified-operators-v5tr5\" (UID: \"f8087c3f-5904-4338-8dcf-779cf0dfa81c\") " pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.662297 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8087c3f-5904-4338-8dcf-779cf0dfa81c-utilities\") pod \"certified-operators-v5tr5\" (UID: \"f8087c3f-5904-4338-8dcf-779cf0dfa81c\") " pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.662318 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8087c3f-5904-4338-8dcf-779cf0dfa81c-catalog-content\") pod \"certified-operators-v5tr5\" (UID: \"f8087c3f-5904-4338-8dcf-779cf0dfa81c\") " pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.662362 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:14 crc kubenswrapper[4744]: E1003 16:27:14.662717 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:15.162704657 +0000 UTC m=+161.442580553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.710713 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.735163 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.765459 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:14 crc kubenswrapper[4744]: E1003 16:27:14.765791 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:15.265772137 +0000 UTC m=+161.545648033 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.765835 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8087c3f-5904-4338-8dcf-779cf0dfa81c-utilities\") pod \"certified-operators-v5tr5\" (UID: \"f8087c3f-5904-4338-8dcf-779cf0dfa81c\") " pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.765859 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8087c3f-5904-4338-8dcf-779cf0dfa81c-catalog-content\") pod \"certified-operators-v5tr5\" (UID: \"f8087c3f-5904-4338-8dcf-779cf0dfa81c\") " pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.765923 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.765961 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vqzf\" (UniqueName: \"kubernetes.io/projected/f8087c3f-5904-4338-8dcf-779cf0dfa81c-kube-api-access-7vqzf\") pod \"certified-operators-v5tr5\" (UID: \"f8087c3f-5904-4338-8dcf-779cf0dfa81c\") " pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.766716 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8087c3f-5904-4338-8dcf-779cf0dfa81c-utilities\") pod \"certified-operators-v5tr5\" (UID: \"f8087c3f-5904-4338-8dcf-779cf0dfa81c\") " pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.766926 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8087c3f-5904-4338-8dcf-779cf0dfa81c-catalog-content\") pod \"certified-operators-v5tr5\" (UID: \"f8087c3f-5904-4338-8dcf-779cf0dfa81c\") " pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:27:14 crc kubenswrapper[4744]: E1003 16:27:14.767180 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:15.267173825 +0000 UTC m=+161.547049721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.796560 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vqzf\" (UniqueName: \"kubernetes.io/projected/f8087c3f-5904-4338-8dcf-779cf0dfa81c-kube-api-access-7vqzf\") pod \"certified-operators-v5tr5\" (UID: \"f8087c3f-5904-4338-8dcf-779cf0dfa81c\") " pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.856297 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7vn7c"] Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.868961 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:14 crc kubenswrapper[4744]: E1003 16:27:14.869484 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:15.369463183 +0000 UTC m=+161.649339079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.874916 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:27:14 crc kubenswrapper[4744]: I1003 16:27:14.970407 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:14 crc kubenswrapper[4744]: E1003 16:27:14.970792 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 16:27:15.470778915 +0000 UTC m=+161.750654811 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6tzpn" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.002786 4744 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.071661 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:15 crc kubenswrapper[4744]: E1003 16:27:15.072132 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 16:27:15.572110937 +0000 UTC m=+161.851986833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.072612 4744 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-03T16:27:15.003181881Z","Handler":null,"Name":""} Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.146712 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nh227"] Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.152744 4744 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.152788 4744 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.173205 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.178433 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.178517 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.225654 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6tzpn\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:15 crc kubenswrapper[4744]: W1003 16:27:15.230157 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18d30e6e_1a94_4d75_8b09_e01c3eb2575e.slice/crio-3cdfd94c609ea4e3994bbd3f1dd85b6f20c5e21b6c3fd175e82bbb5b6598fb8e WatchSource:0}: Error finding container 3cdfd94c609ea4e3994bbd3f1dd85b6f20c5e21b6c3fd175e82bbb5b6598fb8e: Status 404 returned error can't find the container with id 3cdfd94c609ea4e3994bbd3f1dd85b6f20c5e21b6c3fd175e82bbb5b6598fb8e Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.269214 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:15 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:15 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:15 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.269852 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.274954 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.284373 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.301786 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v5tr5"] Oct 03 16:27:15 crc kubenswrapper[4744]: W1003 16:27:15.327298 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8087c3f_5904_4338_8dcf_779cf0dfa81c.slice/crio-9433a4d48c3463a4e9699e6eb670093d66ba96d39aff7607e5c2ee78b41f10df WatchSource:0}: Error finding container 9433a4d48c3463a4e9699e6eb670093d66ba96d39aff7607e5c2ee78b41f10df: Status 404 returned error can't find the container with id 9433a4d48c3463a4e9699e6eb670093d66ba96d39aff7607e5c2ee78b41f10df Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.336520 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mv2tt"] Oct 03 16:27:15 crc kubenswrapper[4744]: W1003 16:27:15.362196 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8be1fdf2_9a41_455e_8315_459cca9f8a09.slice/crio-a6fa806a84299aa83f1b7ce09e33063ad95ff3ccab2967ff0dea127b50a813a5 WatchSource:0}: Error finding container a6fa806a84299aa83f1b7ce09e33063ad95ff3ccab2967ff0dea127b50a813a5: Status 404 returned error can't find the container with id a6fa806a84299aa83f1b7ce09e33063ad95ff3ccab2967ff0dea127b50a813a5 Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.365330 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.385734 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" event={"ID":"36b56a51-ac0b-437b-8f4f-e42476ed6ddd","Type":"ContainerStarted","Data":"f61480f788417ec73d81bd0ceea05c7deb573782568cdfc7df866fd4b33a60cb"} Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.387960 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5tr5" event={"ID":"f8087c3f-5904-4338-8dcf-779cf0dfa81c","Type":"ContainerStarted","Data":"9433a4d48c3463a4e9699e6eb670093d66ba96d39aff7607e5c2ee78b41f10df"} Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.389590 4744 generic.go:334] "Generic (PLEG): container finished" podID="7570f2af-fefc-4726-a9ec-1c27d7cefe43" containerID="3a3c46c528827bacdb2e2e655308cc5eab31fb776a07a5e776e175d2424695d3" exitCode=0 Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.389654 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vn7c" event={"ID":"7570f2af-fefc-4726-a9ec-1c27d7cefe43","Type":"ContainerDied","Data":"3a3c46c528827bacdb2e2e655308cc5eab31fb776a07a5e776e175d2424695d3"} Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.389678 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vn7c" event={"ID":"7570f2af-fefc-4726-a9ec-1c27d7cefe43","Type":"ContainerStarted","Data":"4ba2119cb5dea1c393e5224d33213d0f30614a543f0afd687299d1b428b33c15"} Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.395225 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.395375 4744 generic.go:334] "Generic (PLEG): container finished" podID="62497d3a-2816-455c-824e-d908c11e71e1" containerID="0db7190931f983ae9e46224d23780971bab6bbee1025822a6c21787830fb70a4" exitCode=0 Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.395437 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" event={"ID":"62497d3a-2816-455c-824e-d908c11e71e1","Type":"ContainerDied","Data":"0db7190931f983ae9e46224d23780971bab6bbee1025822a6c21787830fb70a4"} Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.405765 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nh227" event={"ID":"18d30e6e-1a94-4d75-8b09-e01c3eb2575e","Type":"ContainerStarted","Data":"3cdfd94c609ea4e3994bbd3f1dd85b6f20c5e21b6c3fd175e82bbb5b6598fb8e"} Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.651033 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-6tzpn"] Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.955839 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z2cdj"] Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.957426 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.959328 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.977900 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2cdj"] Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.991097 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-utilities\") pod \"redhat-marketplace-z2cdj\" (UID: \"09ab4fbf-98db-4f41-8155-19a9c7cab1ef\") " pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.991604 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npsjm\" (UniqueName: \"kubernetes.io/projected/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-kube-api-access-npsjm\") pod \"redhat-marketplace-z2cdj\" (UID: \"09ab4fbf-98db-4f41-8155-19a9c7cab1ef\") " pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:27:15 crc kubenswrapper[4744]: I1003 16:27:15.991664 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-catalog-content\") pod \"redhat-marketplace-z2cdj\" (UID: \"09ab4fbf-98db-4f41-8155-19a9c7cab1ef\") " pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.093429 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npsjm\" (UniqueName: \"kubernetes.io/projected/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-kube-api-access-npsjm\") pod \"redhat-marketplace-z2cdj\" (UID: \"09ab4fbf-98db-4f41-8155-19a9c7cab1ef\") " pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.093951 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-catalog-content\") pod \"redhat-marketplace-z2cdj\" (UID: \"09ab4fbf-98db-4f41-8155-19a9c7cab1ef\") " pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.094902 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-catalog-content\") pod \"redhat-marketplace-z2cdj\" (UID: \"09ab4fbf-98db-4f41-8155-19a9c7cab1ef\") " pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.094992 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-utilities\") pod \"redhat-marketplace-z2cdj\" (UID: \"09ab4fbf-98db-4f41-8155-19a9c7cab1ef\") " pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.095214 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-utilities\") pod \"redhat-marketplace-z2cdj\" (UID: \"09ab4fbf-98db-4f41-8155-19a9c7cab1ef\") " pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.114014 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npsjm\" (UniqueName: \"kubernetes.io/projected/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-kube-api-access-npsjm\") pod \"redhat-marketplace-z2cdj\" (UID: \"09ab4fbf-98db-4f41-8155-19a9c7cab1ef\") " pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.264483 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:16 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:16 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:16 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.264564 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.274991 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.317415 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lsw7x"] Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.318632 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.330576 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lsw7x"] Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.376592 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jf2dh" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.399872 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-utilities\") pod \"redhat-marketplace-lsw7x\" (UID: \"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056\") " pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.399938 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sssg\" (UniqueName: \"kubernetes.io/projected/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-kube-api-access-8sssg\") pod \"redhat-marketplace-lsw7x\" (UID: \"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056\") " pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.399968 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-catalog-content\") pod \"redhat-marketplace-lsw7x\" (UID: \"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056\") " pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.424058 4744 generic.go:334] "Generic (PLEG): container finished" podID="18d30e6e-1a94-4d75-8b09-e01c3eb2575e" containerID="b0d9eab7d606f87bc36239bcb43357a78e098dfa00b4952f6884c6e47e5d9edb" exitCode=0 Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.424137 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nh227" event={"ID":"18d30e6e-1a94-4d75-8b09-e01c3eb2575e","Type":"ContainerDied","Data":"b0d9eab7d606f87bc36239bcb43357a78e098dfa00b4952f6884c6e47e5d9edb"} Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.426518 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" event={"ID":"cc342582-3925-4af1-b9b8-8e1333af091c","Type":"ContainerStarted","Data":"b9e2148bb457a4bd1e40b0a8cd550f2d60b5d1bb008524d78e75de3b1cd3550d"} Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.426587 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" event={"ID":"cc342582-3925-4af1-b9b8-8e1333af091c","Type":"ContainerStarted","Data":"f4bb15b9977ff10a5b109cde5d357b0684eec284957bd238e917118e253694b5"} Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.426794 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.449663 4744 generic.go:334] "Generic (PLEG): container finished" podID="f8087c3f-5904-4338-8dcf-779cf0dfa81c" containerID="18897b7c8387301650528ece35085546a8d97cdd2d8f99292d1a122540497e92" exitCode=0 Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.450012 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5tr5" event={"ID":"f8087c3f-5904-4338-8dcf-779cf0dfa81c","Type":"ContainerDied","Data":"18897b7c8387301650528ece35085546a8d97cdd2d8f99292d1a122540497e92"} Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.456458 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" event={"ID":"36b56a51-ac0b-437b-8f4f-e42476ed6ddd","Type":"ContainerStarted","Data":"b8f2e74a628a578e9d6e75722bbe54961bfd25298969b5874710630339548580"} Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.469308 4744 generic.go:334] "Generic (PLEG): container finished" podID="8be1fdf2-9a41-455e-8315-459cca9f8a09" containerID="ae2f5ecd5917dd71ddd27e612b5a2df04b7d330a31e8722d293f00b97ddad8a3" exitCode=0 Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.470317 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mv2tt" event={"ID":"8be1fdf2-9a41-455e-8315-459cca9f8a09","Type":"ContainerDied","Data":"ae2f5ecd5917dd71ddd27e612b5a2df04b7d330a31e8722d293f00b97ddad8a3"} Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.471136 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mv2tt" event={"ID":"8be1fdf2-9a41-455e-8315-459cca9f8a09","Type":"ContainerStarted","Data":"a6fa806a84299aa83f1b7ce09e33063ad95ff3ccab2967ff0dea127b50a813a5"} Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.477041 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" podStartSLOduration=140.477020866 podStartE2EDuration="2m20.477020866s" podCreationTimestamp="2025-10-03 16:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:16.467809505 +0000 UTC m=+162.747685401" watchObservedRunningTime="2025-10-03 16:27:16.477020866 +0000 UTC m=+162.756896762" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.501274 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sssg\" (UniqueName: \"kubernetes.io/projected/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-kube-api-access-8sssg\") pod \"redhat-marketplace-lsw7x\" (UID: \"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056\") " pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.501364 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-catalog-content\") pod \"redhat-marketplace-lsw7x\" (UID: \"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056\") " pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.504379 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-catalog-content\") pod \"redhat-marketplace-lsw7x\" (UID: \"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056\") " pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.506662 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-d6zzl" podStartSLOduration=12.506640127 podStartE2EDuration="12.506640127s" podCreationTimestamp="2025-10-03 16:27:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:16.498464793 +0000 UTC m=+162.778340689" watchObservedRunningTime="2025-10-03 16:27:16.506640127 +0000 UTC m=+162.786516023" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.509654 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-utilities\") pod \"redhat-marketplace-lsw7x\" (UID: \"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056\") " pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.510712 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-utilities\") pod \"redhat-marketplace-lsw7x\" (UID: \"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056\") " pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.523681 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sssg\" (UniqueName: \"kubernetes.io/projected/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-kube-api-access-8sssg\") pod \"redhat-marketplace-lsw7x\" (UID: \"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056\") " pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.579840 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2cdj"] Oct 03 16:27:16 crc kubenswrapper[4744]: W1003 16:27:16.598307 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09ab4fbf_98db_4f41_8155_19a9c7cab1ef.slice/crio-6441d2c33c5ef0c47bb49f40f445e60acbd0c87fcc1b76926c190471376e156e WatchSource:0}: Error finding container 6441d2c33c5ef0c47bb49f40f445e60acbd0c87fcc1b76926c190471376e156e: Status 404 returned error can't find the container with id 6441d2c33c5ef0c47bb49f40f445e60acbd0c87fcc1b76926c190471376e156e Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.605118 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.646909 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.714236 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.784767 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 03 16:27:16 crc kubenswrapper[4744]: E1003 16:27:16.785661 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62497d3a-2816-455c-824e-d908c11e71e1" containerName="collect-profiles" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.785758 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="62497d3a-2816-455c-824e-d908c11e71e1" containerName="collect-profiles" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.785992 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="62497d3a-2816-455c-824e-d908c11e71e1" containerName="collect-profiles" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.786476 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.792525 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.792908 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.796534 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.814209 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62497d3a-2816-455c-824e-d908c11e71e1-secret-volume\") pod \"62497d3a-2816-455c-824e-d908c11e71e1\" (UID: \"62497d3a-2816-455c-824e-d908c11e71e1\") " Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.814425 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62497d3a-2816-455c-824e-d908c11e71e1-config-volume\") pod \"62497d3a-2816-455c-824e-d908c11e71e1\" (UID: \"62497d3a-2816-455c-824e-d908c11e71e1\") " Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.814517 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwh72\" (UniqueName: \"kubernetes.io/projected/62497d3a-2816-455c-824e-d908c11e71e1-kube-api-access-fwh72\") pod \"62497d3a-2816-455c-824e-d908c11e71e1\" (UID: \"62497d3a-2816-455c-824e-d908c11e71e1\") " Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.814725 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9de21d2-7004-4b21-bf31-ef1c280aa940-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c9de21d2-7004-4b21-bf31-ef1c280aa940\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.814829 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c9de21d2-7004-4b21-bf31-ef1c280aa940-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c9de21d2-7004-4b21-bf31-ef1c280aa940\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.816522 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62497d3a-2816-455c-824e-d908c11e71e1-config-volume" (OuterVolumeSpecName: "config-volume") pod "62497d3a-2816-455c-824e-d908c11e71e1" (UID: "62497d3a-2816-455c-824e-d908c11e71e1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.824130 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62497d3a-2816-455c-824e-d908c11e71e1-kube-api-access-fwh72" (OuterVolumeSpecName: "kube-api-access-fwh72") pod "62497d3a-2816-455c-824e-d908c11e71e1" (UID: "62497d3a-2816-455c-824e-d908c11e71e1"). InnerVolumeSpecName "kube-api-access-fwh72". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.832099 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62497d3a-2816-455c-824e-d908c11e71e1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "62497d3a-2816-455c-824e-d908c11e71e1" (UID: "62497d3a-2816-455c-824e-d908c11e71e1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.906832 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.924052 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9de21d2-7004-4b21-bf31-ef1c280aa940-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c9de21d2-7004-4b21-bf31-ef1c280aa940\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.924161 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c9de21d2-7004-4b21-bf31-ef1c280aa940-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c9de21d2-7004-4b21-bf31-ef1c280aa940\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.924222 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/62497d3a-2816-455c-824e-d908c11e71e1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.924236 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwh72\" (UniqueName: \"kubernetes.io/projected/62497d3a-2816-455c-824e-d908c11e71e1-kube-api-access-fwh72\") on node \"crc\" DevicePath \"\"" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.924250 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/62497d3a-2816-455c-824e-d908c11e71e1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.924307 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c9de21d2-7004-4b21-bf31-ef1c280aa940-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c9de21d2-7004-4b21-bf31-ef1c280aa940\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.944559 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9de21d2-7004-4b21-bf31-ef1c280aa940-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c9de21d2-7004-4b21-bf31-ef1c280aa940\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 16:27:16 crc kubenswrapper[4744]: I1003 16:27:16.972481 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lsw7x"] Oct 03 16:27:16 crc kubenswrapper[4744]: W1003 16:27:16.989695 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bb5b864_3bb2_4e32_9e0a_63f54c1f8056.slice/crio-83037ff1f46bca709bd33a9e26da6e55599d2a4ac8c323370330e393ebd506ea WatchSource:0}: Error finding container 83037ff1f46bca709bd33a9e26da6e55599d2a4ac8c323370330e393ebd506ea: Status 404 returned error can't find the container with id 83037ff1f46bca709bd33a9e26da6e55599d2a4ac8c323370330e393ebd506ea Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.111457 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.115875 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7jvbs"] Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.117360 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.119854 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.139115 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7jvbs"] Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.193836 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-thbg6 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.193893 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-thbg6" podUID="6ac85db3-283c-43f9-9925-b8a0f8354181" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.194181 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-thbg6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.194212 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-thbg6" podUID="6ac85db3-283c-43f9-9925-b8a0f8354181" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.227522 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh2lw\" (UniqueName: \"kubernetes.io/projected/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-kube-api-access-lh2lw\") pod \"redhat-operators-7jvbs\" (UID: \"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b\") " pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.227852 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-utilities\") pod \"redhat-operators-7jvbs\" (UID: \"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b\") " pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.227913 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-catalog-content\") pod \"redhat-operators-7jvbs\" (UID: \"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b\") " pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.238396 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.239019 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.244657 4744 patch_prober.go:28] interesting pod/console-f9d7485db-v5snt container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.244726 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-v5snt" podUID="d61086b0-b78f-4239-8d9a-49aba28c5820" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.261314 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.278722 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:17 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:17 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:17 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.278813 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.330418 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-catalog-content\") pod \"redhat-operators-7jvbs\" (UID: \"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b\") " pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.330702 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh2lw\" (UniqueName: \"kubernetes.io/projected/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-kube-api-access-lh2lw\") pod \"redhat-operators-7jvbs\" (UID: \"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b\") " pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.330858 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-utilities\") pod \"redhat-operators-7jvbs\" (UID: \"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b\") " pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.332751 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-catalog-content\") pod \"redhat-operators-7jvbs\" (UID: \"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b\") " pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.336393 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-utilities\") pod \"redhat-operators-7jvbs\" (UID: \"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b\") " pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.366725 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh2lw\" (UniqueName: \"kubernetes.io/projected/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-kube-api-access-lh2lw\") pod \"redhat-operators-7jvbs\" (UID: \"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b\") " pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.437538 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.500719 4744 generic.go:334] "Generic (PLEG): container finished" podID="09ab4fbf-98db-4f41-8155-19a9c7cab1ef" containerID="afa91b7a2c29f18be0e6cf3b37b460db5808f4ddf381dbf79fbf8b9840deb89e" exitCode=0 Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.500920 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2cdj" event={"ID":"09ab4fbf-98db-4f41-8155-19a9c7cab1ef","Type":"ContainerDied","Data":"afa91b7a2c29f18be0e6cf3b37b460db5808f4ddf381dbf79fbf8b9840deb89e"} Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.500971 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2cdj" event={"ID":"09ab4fbf-98db-4f41-8155-19a9c7cab1ef","Type":"ContainerStarted","Data":"6441d2c33c5ef0c47bb49f40f445e60acbd0c87fcc1b76926c190471376e156e"} Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.507897 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.508237 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv" event={"ID":"62497d3a-2816-455c-824e-d908c11e71e1","Type":"ContainerDied","Data":"f4e8b264982619990bf78dccca62adade613a46d3ef8d825f82d5a3d4779ae2a"} Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.508280 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4e8b264982619990bf78dccca62adade613a46d3ef8d825f82d5a3d4779ae2a" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.520565 4744 generic.go:334] "Generic (PLEG): container finished" podID="1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" containerID="e0bd0e4f690ef101b1e80ded5d26752b2a88b7bd81742bc2f065322d5f5df47b" exitCode=0 Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.520681 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lsw7x" event={"ID":"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056","Type":"ContainerDied","Data":"e0bd0e4f690ef101b1e80ded5d26752b2a88b7bd81742bc2f065322d5f5df47b"} Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.520717 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lsw7x" event={"ID":"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056","Type":"ContainerStarted","Data":"83037ff1f46bca709bd33a9e26da6e55599d2a4ac8c323370330e393ebd506ea"} Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.528308 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.528345 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.530940 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2m27n"] Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.532202 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.544824 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.557417 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2m27n"] Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.570411 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 03 16:27:17 crc kubenswrapper[4744]: W1003 16:27:17.588968 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podc9de21d2_7004_4b21_bf31_ef1c280aa940.slice/crio-35a61c064b221542d4a49415cff84c472cf917a6a927bd0b6cb682945d20962f WatchSource:0}: Error finding container 35a61c064b221542d4a49415cff84c472cf917a6a927bd0b6cb682945d20962f: Status 404 returned error can't find the container with id 35a61c064b221542d4a49415cff84c472cf917a6a927bd0b6cb682945d20962f Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.635690 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39443df9-cc38-4b33-b02e-e9136e902fc2-utilities\") pod \"redhat-operators-2m27n\" (UID: \"39443df9-cc38-4b33-b02e-e9136e902fc2\") " pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.636064 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39443df9-cc38-4b33-b02e-e9136e902fc2-catalog-content\") pod \"redhat-operators-2m27n\" (UID: \"39443df9-cc38-4b33-b02e-e9136e902fc2\") " pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.636116 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2vrm\" (UniqueName: \"kubernetes.io/projected/39443df9-cc38-4b33-b02e-e9136e902fc2-kube-api-access-v2vrm\") pod \"redhat-operators-2m27n\" (UID: \"39443df9-cc38-4b33-b02e-e9136e902fc2\") " pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.737948 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39443df9-cc38-4b33-b02e-e9136e902fc2-catalog-content\") pod \"redhat-operators-2m27n\" (UID: \"39443df9-cc38-4b33-b02e-e9136e902fc2\") " pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.738408 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2vrm\" (UniqueName: \"kubernetes.io/projected/39443df9-cc38-4b33-b02e-e9136e902fc2-kube-api-access-v2vrm\") pod \"redhat-operators-2m27n\" (UID: \"39443df9-cc38-4b33-b02e-e9136e902fc2\") " pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.738508 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39443df9-cc38-4b33-b02e-e9136e902fc2-utilities\") pod \"redhat-operators-2m27n\" (UID: \"39443df9-cc38-4b33-b02e-e9136e902fc2\") " pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.738834 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39443df9-cc38-4b33-b02e-e9136e902fc2-catalog-content\") pod \"redhat-operators-2m27n\" (UID: \"39443df9-cc38-4b33-b02e-e9136e902fc2\") " pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.738950 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39443df9-cc38-4b33-b02e-e9136e902fc2-utilities\") pod \"redhat-operators-2m27n\" (UID: \"39443df9-cc38-4b33-b02e-e9136e902fc2\") " pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.773291 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2vrm\" (UniqueName: \"kubernetes.io/projected/39443df9-cc38-4b33-b02e-e9136e902fc2-kube-api-access-v2vrm\") pod \"redhat-operators-2m27n\" (UID: \"39443df9-cc38-4b33-b02e-e9136e902fc2\") " pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.847970 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7jvbs"] Oct 03 16:27:17 crc kubenswrapper[4744]: I1003 16:27:17.857381 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:27:17 crc kubenswrapper[4744]: W1003 16:27:17.888386 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b5a6267_63e7_44e7_a891_2ab9c9b8b29b.slice/crio-82eb980126a1a1f312e22a9a12c87a2520b46d9fb5dee253218bfe2940f4b268 WatchSource:0}: Error finding container 82eb980126a1a1f312e22a9a12c87a2520b46d9fb5dee253218bfe2940f4b268: Status 404 returned error can't find the container with id 82eb980126a1a1f312e22a9a12c87a2520b46d9fb5dee253218bfe2940f4b268 Oct 03 16:27:18 crc kubenswrapper[4744]: I1003 16:27:18.018052 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" Oct 03 16:27:18 crc kubenswrapper[4744]: I1003 16:27:18.048365 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs\") pod \"network-metrics-daemon-4f7dk\" (UID: \"1008bf15-63d7-45f4-b42b-a4680b7dc232\") " pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:27:18 crc kubenswrapper[4744]: I1003 16:27:18.059681 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1008bf15-63d7-45f4-b42b-a4680b7dc232-metrics-certs\") pod \"network-metrics-daemon-4f7dk\" (UID: \"1008bf15-63d7-45f4-b42b-a4680b7dc232\") " pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:27:18 crc kubenswrapper[4744]: I1003 16:27:18.264605 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:18 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:18 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:18 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:18 crc kubenswrapper[4744]: I1003 16:27:18.265141 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:18 crc kubenswrapper[4744]: I1003 16:27:18.308272 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4f7dk" Oct 03 16:27:18 crc kubenswrapper[4744]: I1003 16:27:18.441395 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2m27n"] Oct 03 16:27:18 crc kubenswrapper[4744]: W1003 16:27:18.463947 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39443df9_cc38_4b33_b02e_e9136e902fc2.slice/crio-9e928f51ec30ac785424065feb0ff395379a63c6204db8ce5305dd09bbf3f051 WatchSource:0}: Error finding container 9e928f51ec30ac785424065feb0ff395379a63c6204db8ce5305dd09bbf3f051: Status 404 returned error can't find the container with id 9e928f51ec30ac785424065feb0ff395379a63c6204db8ce5305dd09bbf3f051 Oct 03 16:27:18 crc kubenswrapper[4744]: I1003 16:27:18.530908 4744 generic.go:334] "Generic (PLEG): container finished" podID="7b5a6267-63e7-44e7-a891-2ab9c9b8b29b" containerID="035308c2a9ef1e069c6af81885bcad9b0bed87952ee1d25c2fd87e08ab65ecca" exitCode=0 Oct 03 16:27:18 crc kubenswrapper[4744]: I1003 16:27:18.531017 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jvbs" event={"ID":"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b","Type":"ContainerDied","Data":"035308c2a9ef1e069c6af81885bcad9b0bed87952ee1d25c2fd87e08ab65ecca"} Oct 03 16:27:18 crc kubenswrapper[4744]: I1003 16:27:18.531060 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jvbs" event={"ID":"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b","Type":"ContainerStarted","Data":"82eb980126a1a1f312e22a9a12c87a2520b46d9fb5dee253218bfe2940f4b268"} Oct 03 16:27:18 crc kubenswrapper[4744]: I1003 16:27:18.540746 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2m27n" event={"ID":"39443df9-cc38-4b33-b02e-e9136e902fc2","Type":"ContainerStarted","Data":"9e928f51ec30ac785424065feb0ff395379a63c6204db8ce5305dd09bbf3f051"} Oct 03 16:27:18 crc kubenswrapper[4744]: I1003 16:27:18.558915 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c9de21d2-7004-4b21-bf31-ef1c280aa940","Type":"ContainerStarted","Data":"35a61c064b221542d4a49415cff84c472cf917a6a927bd0b6cb682945d20962f"} Oct 03 16:27:18 crc kubenswrapper[4744]: I1003 16:27:18.566676 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-2pjgv" Oct 03 16:27:18 crc kubenswrapper[4744]: I1003 16:27:18.586112 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.586055709 podStartE2EDuration="2.586055709s" podCreationTimestamp="2025-10-03 16:27:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:18.582799719 +0000 UTC m=+164.862675615" watchObservedRunningTime="2025-10-03 16:27:18.586055709 +0000 UTC m=+164.865931605" Oct 03 16:27:18 crc kubenswrapper[4744]: I1003 16:27:18.997395 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4f7dk"] Oct 03 16:27:19 crc kubenswrapper[4744]: W1003 16:27:19.042990 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1008bf15_63d7_45f4_b42b_a4680b7dc232.slice/crio-06b0e42d6663d02b54f449df1f81172e902a89dcbbee109a3fe585acfd6200ee WatchSource:0}: Error finding container 06b0e42d6663d02b54f449df1f81172e902a89dcbbee109a3fe585acfd6200ee: Status 404 returned error can't find the container with id 06b0e42d6663d02b54f449df1f81172e902a89dcbbee109a3fe585acfd6200ee Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.168830 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.170077 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.175395 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.175733 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.184385 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.193382 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ee70474c-bce0-4289-af10-7e1ae1393b5a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ee70474c-bce0-4289-af10-7e1ae1393b5a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.193437 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee70474c-bce0-4289-af10-7e1ae1393b5a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ee70474c-bce0-4289-af10-7e1ae1393b5a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.264726 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:19 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:19 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:19 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.264792 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.297004 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ee70474c-bce0-4289-af10-7e1ae1393b5a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ee70474c-bce0-4289-af10-7e1ae1393b5a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.297076 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee70474c-bce0-4289-af10-7e1ae1393b5a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ee70474c-bce0-4289-af10-7e1ae1393b5a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.297464 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ee70474c-bce0-4289-af10-7e1ae1393b5a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ee70474c-bce0-4289-af10-7e1ae1393b5a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.324398 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee70474c-bce0-4289-af10-7e1ae1393b5a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ee70474c-bce0-4289-af10-7e1ae1393b5a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.498800 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.627948 4744 generic.go:334] "Generic (PLEG): container finished" podID="39443df9-cc38-4b33-b02e-e9136e902fc2" containerID="5f0fe5f98666cf3af5a95ac6585c9546bb04e5e02f564de4332efccb264d6dfc" exitCode=0 Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.628058 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2m27n" event={"ID":"39443df9-cc38-4b33-b02e-e9136e902fc2","Type":"ContainerDied","Data":"5f0fe5f98666cf3af5a95ac6585c9546bb04e5e02f564de4332efccb264d6dfc"} Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.635126 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" event={"ID":"1008bf15-63d7-45f4-b42b-a4680b7dc232","Type":"ContainerStarted","Data":"06b0e42d6663d02b54f449df1f81172e902a89dcbbee109a3fe585acfd6200ee"} Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.641163 4744 generic.go:334] "Generic (PLEG): container finished" podID="c9de21d2-7004-4b21-bf31-ef1c280aa940" containerID="313118b7265a2d404f448e5fe72eddbf559255a8144829f2d364a878d847e886" exitCode=0 Oct 03 16:27:19 crc kubenswrapper[4744]: I1003 16:27:19.641734 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c9de21d2-7004-4b21-bf31-ef1c280aa940","Type":"ContainerDied","Data":"313118b7265a2d404f448e5fe72eddbf559255a8144829f2d364a878d847e886"} Oct 03 16:27:20 crc kubenswrapper[4744]: I1003 16:27:20.104139 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 03 16:27:20 crc kubenswrapper[4744]: I1003 16:27:20.278471 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:20 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:20 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:20 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:20 crc kubenswrapper[4744]: I1003 16:27:20.279003 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:20 crc kubenswrapper[4744]: I1003 16:27:20.825396 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" event={"ID":"1008bf15-63d7-45f4-b42b-a4680b7dc232","Type":"ContainerStarted","Data":"3645408d7c57c661571160f4a4d0e6c7980aea688f6018f89487b3ed03a7af5c"} Oct 03 16:27:20 crc kubenswrapper[4744]: I1003 16:27:20.831043 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ee70474c-bce0-4289-af10-7e1ae1393b5a","Type":"ContainerStarted","Data":"b36c9f478d2d2f4614bebe1373c802335e88926c36f808ae9d920b3b641904c5"} Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.266471 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:21 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:21 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:21 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.266545 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.342407 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.464679 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c9de21d2-7004-4b21-bf31-ef1c280aa940-kubelet-dir\") pod \"c9de21d2-7004-4b21-bf31-ef1c280aa940\" (UID: \"c9de21d2-7004-4b21-bf31-ef1c280aa940\") " Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.464795 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9de21d2-7004-4b21-bf31-ef1c280aa940-kube-api-access\") pod \"c9de21d2-7004-4b21-bf31-ef1c280aa940\" (UID: \"c9de21d2-7004-4b21-bf31-ef1c280aa940\") " Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.464822 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c9de21d2-7004-4b21-bf31-ef1c280aa940-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c9de21d2-7004-4b21-bf31-ef1c280aa940" (UID: "c9de21d2-7004-4b21-bf31-ef1c280aa940"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.465370 4744 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c9de21d2-7004-4b21-bf31-ef1c280aa940-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.485958 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9de21d2-7004-4b21-bf31-ef1c280aa940-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c9de21d2-7004-4b21-bf31-ef1c280aa940" (UID: "c9de21d2-7004-4b21-bf31-ef1c280aa940"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.566856 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9de21d2-7004-4b21-bf31-ef1c280aa940-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.855251 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.855294 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c9de21d2-7004-4b21-bf31-ef1c280aa940","Type":"ContainerDied","Data":"35a61c064b221542d4a49415cff84c472cf917a6a927bd0b6cb682945d20962f"} Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.855354 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35a61c064b221542d4a49415cff84c472cf917a6a927bd0b6cb682945d20962f" Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.889755 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4f7dk" event={"ID":"1008bf15-63d7-45f4-b42b-a4680b7dc232","Type":"ContainerStarted","Data":"a64c02c7d55b44eb6ca48436a18e0b22ed04fb930e964bac6a48beb73b69f160"} Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.907479 4744 generic.go:334] "Generic (PLEG): container finished" podID="ee70474c-bce0-4289-af10-7e1ae1393b5a" containerID="4f42f1c7d877ba3124c93a2421a79524db774662331fc2e7ed9e2d5e53a63d60" exitCode=0 Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.907546 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ee70474c-bce0-4289-af10-7e1ae1393b5a","Type":"ContainerDied","Data":"4f42f1c7d877ba3124c93a2421a79524db774662331fc2e7ed9e2d5e53a63d60"} Oct 03 16:27:21 crc kubenswrapper[4744]: I1003 16:27:21.920344 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-4f7dk" podStartSLOduration=146.920312745 podStartE2EDuration="2m26.920312745s" podCreationTimestamp="2025-10-03 16:24:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:27:21.915863184 +0000 UTC m=+168.195739080" watchObservedRunningTime="2025-10-03 16:27:21.920312745 +0000 UTC m=+168.200188651" Oct 03 16:27:22 crc kubenswrapper[4744]: I1003 16:27:22.263472 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:22 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:22 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:22 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:22 crc kubenswrapper[4744]: I1003 16:27:22.263547 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:23 crc kubenswrapper[4744]: I1003 16:27:23.081183 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-qmddz" Oct 03 16:27:23 crc kubenswrapper[4744]: I1003 16:27:23.275928 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:23 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:23 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:23 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:23 crc kubenswrapper[4744]: I1003 16:27:23.276382 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:23 crc kubenswrapper[4744]: I1003 16:27:23.434427 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 16:27:23 crc kubenswrapper[4744]: I1003 16:27:23.607692 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee70474c-bce0-4289-af10-7e1ae1393b5a-kube-api-access\") pod \"ee70474c-bce0-4289-af10-7e1ae1393b5a\" (UID: \"ee70474c-bce0-4289-af10-7e1ae1393b5a\") " Oct 03 16:27:23 crc kubenswrapper[4744]: I1003 16:27:23.607892 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ee70474c-bce0-4289-af10-7e1ae1393b5a-kubelet-dir\") pod \"ee70474c-bce0-4289-af10-7e1ae1393b5a\" (UID: \"ee70474c-bce0-4289-af10-7e1ae1393b5a\") " Oct 03 16:27:23 crc kubenswrapper[4744]: I1003 16:27:23.608055 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee70474c-bce0-4289-af10-7e1ae1393b5a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ee70474c-bce0-4289-af10-7e1ae1393b5a" (UID: "ee70474c-bce0-4289-af10-7e1ae1393b5a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:27:23 crc kubenswrapper[4744]: I1003 16:27:23.608323 4744 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ee70474c-bce0-4289-af10-7e1ae1393b5a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 03 16:27:23 crc kubenswrapper[4744]: I1003 16:27:23.616734 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee70474c-bce0-4289-af10-7e1ae1393b5a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ee70474c-bce0-4289-af10-7e1ae1393b5a" (UID: "ee70474c-bce0-4289-af10-7e1ae1393b5a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:27:23 crc kubenswrapper[4744]: I1003 16:27:23.710019 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee70474c-bce0-4289-af10-7e1ae1393b5a-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 16:27:23 crc kubenswrapper[4744]: I1003 16:27:23.940956 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ee70474c-bce0-4289-af10-7e1ae1393b5a","Type":"ContainerDied","Data":"b36c9f478d2d2f4614bebe1373c802335e88926c36f808ae9d920b3b641904c5"} Oct 03 16:27:23 crc kubenswrapper[4744]: I1003 16:27:23.941007 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b36c9f478d2d2f4614bebe1373c802335e88926c36f808ae9d920b3b641904c5" Oct 03 16:27:23 crc kubenswrapper[4744]: I1003 16:27:23.941009 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 16:27:24 crc kubenswrapper[4744]: I1003 16:27:24.269933 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:24 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:24 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:24 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:24 crc kubenswrapper[4744]: I1003 16:27:24.270004 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:25 crc kubenswrapper[4744]: I1003 16:27:25.264632 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:25 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:25 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:25 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:25 crc kubenswrapper[4744]: I1003 16:27:25.265005 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:26 crc kubenswrapper[4744]: I1003 16:27:26.263566 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:26 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Oct 03 16:27:26 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:26 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:26 crc kubenswrapper[4744]: I1003 16:27:26.263650 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:27 crc kubenswrapper[4744]: I1003 16:27:27.198635 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-thbg6" Oct 03 16:27:27 crc kubenswrapper[4744]: I1003 16:27:27.238555 4744 patch_prober.go:28] interesting pod/console-f9d7485db-v5snt container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Oct 03 16:27:27 crc kubenswrapper[4744]: I1003 16:27:27.238612 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-v5snt" podUID="d61086b0-b78f-4239-8d9a-49aba28c5820" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Oct 03 16:27:27 crc kubenswrapper[4744]: I1003 16:27:27.264311 4744 patch_prober.go:28] interesting pod/router-default-5444994796-946kz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 16:27:27 crc kubenswrapper[4744]: [+]has-synced ok Oct 03 16:27:27 crc kubenswrapper[4744]: [+]process-running ok Oct 03 16:27:27 crc kubenswrapper[4744]: healthz check failed Oct 03 16:27:27 crc kubenswrapper[4744]: I1003 16:27:27.264392 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-946kz" podUID="b550c60f-b9e0-44a5-938f-de24afa6c123" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:27:28 crc kubenswrapper[4744]: I1003 16:27:28.263908 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:28 crc kubenswrapper[4744]: I1003 16:27:28.266809 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-946kz" Oct 03 16:27:33 crc kubenswrapper[4744]: I1003 16:27:33.220204 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 16:27:34 crc kubenswrapper[4744]: I1003 16:27:34.668970 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:27:34 crc kubenswrapper[4744]: I1003 16:27:34.669665 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:27:35 crc kubenswrapper[4744]: I1003 16:27:35.372631 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:27:37 crc kubenswrapper[4744]: I1003 16:27:37.535728 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:37 crc kubenswrapper[4744]: I1003 16:27:37.541678 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:27:48 crc kubenswrapper[4744]: I1003 16:27:48.026277 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vqdgh" Oct 03 16:27:50 crc kubenswrapper[4744]: E1003 16:27:50.496283 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 03 16:27:50 crc kubenswrapper[4744]: E1003 16:27:50.496932 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hk9qh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-nh227_openshift-marketplace(18d30e6e-1a94-4d75-8b09-e01c3eb2575e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 16:27:50 crc kubenswrapper[4744]: E1003 16:27:50.498129 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-nh227" podUID="18d30e6e-1a94-4d75-8b09-e01c3eb2575e" Oct 03 16:27:50 crc kubenswrapper[4744]: E1003 16:27:50.532842 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 03 16:27:50 crc kubenswrapper[4744]: E1003 16:27:50.533060 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7vqzf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-v5tr5_openshift-marketplace(f8087c3f-5904-4338-8dcf-779cf0dfa81c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 16:27:50 crc kubenswrapper[4744]: E1003 16:27:50.535392 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-v5tr5" podUID="f8087c3f-5904-4338-8dcf-779cf0dfa81c" Oct 03 16:27:50 crc kubenswrapper[4744]: E1003 16:27:50.593579 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 03 16:27:50 crc kubenswrapper[4744]: E1003 16:27:50.593883 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v2vrm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-2m27n_openshift-marketplace(39443df9-cc38-4b33-b02e-e9136e902fc2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 16:27:50 crc kubenswrapper[4744]: E1003 16:27:50.595962 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-2m27n" podUID="39443df9-cc38-4b33-b02e-e9136e902fc2" Oct 03 16:27:52 crc kubenswrapper[4744]: E1003 16:27:52.270854 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-v5tr5" podUID="f8087c3f-5904-4338-8dcf-779cf0dfa81c" Oct 03 16:27:52 crc kubenswrapper[4744]: E1003 16:27:52.271032 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-2m27n" podUID="39443df9-cc38-4b33-b02e-e9136e902fc2" Oct 03 16:27:52 crc kubenswrapper[4744]: E1003 16:27:52.271218 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-nh227" podUID="18d30e6e-1a94-4d75-8b09-e01c3eb2575e" Oct 03 16:27:52 crc kubenswrapper[4744]: E1003 16:27:52.349754 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 03 16:27:52 crc kubenswrapper[4744]: E1003 16:27:52.350000 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lxm8k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-mv2tt_openshift-marketplace(8be1fdf2-9a41-455e-8315-459cca9f8a09): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 16:27:52 crc kubenswrapper[4744]: E1003 16:27:52.370163 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 03 16:27:52 crc kubenswrapper[4744]: E1003 16:27:52.380977 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kpl6l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-7vn7c_openshift-marketplace(7570f2af-fefc-4726-a9ec-1c27d7cefe43): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 16:27:52 crc kubenswrapper[4744]: E1003 16:27:52.370370 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-mv2tt" podUID="8be1fdf2-9a41-455e-8315-459cca9f8a09" Oct 03 16:27:52 crc kubenswrapper[4744]: E1003 16:27:52.382105 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-7vn7c" podUID="7570f2af-fefc-4726-a9ec-1c27d7cefe43" Oct 03 16:27:55 crc kubenswrapper[4744]: E1003 16:27:55.022251 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-7vn7c" podUID="7570f2af-fefc-4726-a9ec-1c27d7cefe43" Oct 03 16:27:55 crc kubenswrapper[4744]: E1003 16:27:55.022369 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-mv2tt" podUID="8be1fdf2-9a41-455e-8315-459cca9f8a09" Oct 03 16:27:55 crc kubenswrapper[4744]: E1003 16:27:55.741361 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 03 16:27:55 crc kubenswrapper[4744]: E1003 16:27:55.742114 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8sssg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-lsw7x_openshift-marketplace(1bb5b864-3bb2-4e32-9e0a-63f54c1f8056): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 16:27:55 crc kubenswrapper[4744]: E1003 16:27:55.743379 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-lsw7x" podUID="1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" Oct 03 16:27:55 crc kubenswrapper[4744]: E1003 16:27:55.757084 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 03 16:27:55 crc kubenswrapper[4744]: E1003 16:27:55.757288 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-npsjm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-z2cdj_openshift-marketplace(09ab4fbf-98db-4f41-8155-19a9c7cab1ef): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 16:27:55 crc kubenswrapper[4744]: E1003 16:27:55.758715 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-z2cdj" podUID="09ab4fbf-98db-4f41-8155-19a9c7cab1ef" Oct 03 16:27:56 crc kubenswrapper[4744]: I1003 16:27:56.158726 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jvbs" event={"ID":"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b","Type":"ContainerStarted","Data":"09822b10139977124afca9bc3e46b3c5d9af0eaa3d2509c4f60914084e0c7e2d"} Oct 03 16:27:56 crc kubenswrapper[4744]: E1003 16:27:56.161714 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-lsw7x" podUID="1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" Oct 03 16:27:56 crc kubenswrapper[4744]: E1003 16:27:56.161714 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-z2cdj" podUID="09ab4fbf-98db-4f41-8155-19a9c7cab1ef" Oct 03 16:27:57 crc kubenswrapper[4744]: I1003 16:27:57.170740 4744 generic.go:334] "Generic (PLEG): container finished" podID="7b5a6267-63e7-44e7-a891-2ab9c9b8b29b" containerID="09822b10139977124afca9bc3e46b3c5d9af0eaa3d2509c4f60914084e0c7e2d" exitCode=0 Oct 03 16:27:57 crc kubenswrapper[4744]: I1003 16:27:57.170850 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jvbs" event={"ID":"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b","Type":"ContainerDied","Data":"09822b10139977124afca9bc3e46b3c5d9af0eaa3d2509c4f60914084e0c7e2d"} Oct 03 16:27:58 crc kubenswrapper[4744]: I1003 16:27:58.185256 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jvbs" event={"ID":"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b","Type":"ContainerStarted","Data":"9c31a04e5a9f04d92476c63fc866a57202ae33eb8a3a996e498f4294cf616bfb"} Oct 03 16:27:58 crc kubenswrapper[4744]: I1003 16:27:58.210457 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7jvbs" podStartSLOduration=2.048640733 podStartE2EDuration="41.210434161s" podCreationTimestamp="2025-10-03 16:27:17 +0000 UTC" firstStartedPulling="2025-10-03 16:27:18.533676816 +0000 UTC m=+164.813552712" lastFinishedPulling="2025-10-03 16:27:57.695470244 +0000 UTC m=+203.975346140" observedRunningTime="2025-10-03 16:27:58.208443496 +0000 UTC m=+204.488319392" watchObservedRunningTime="2025-10-03 16:27:58.210434161 +0000 UTC m=+204.490310057" Oct 03 16:28:04 crc kubenswrapper[4744]: I1003 16:28:04.668438 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:28:04 crc kubenswrapper[4744]: I1003 16:28:04.669581 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:28:04 crc kubenswrapper[4744]: I1003 16:28:04.669686 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:28:04 crc kubenswrapper[4744]: I1003 16:28:04.671009 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 16:28:04 crc kubenswrapper[4744]: I1003 16:28:04.671196 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e" gracePeriod=600 Oct 03 16:28:05 crc kubenswrapper[4744]: I1003 16:28:05.223379 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2m27n" event={"ID":"39443df9-cc38-4b33-b02e-e9136e902fc2","Type":"ContainerStarted","Data":"338cffb6c214402baef247496030860afa44ee0247f83ba4fadcab1226312d9d"} Oct 03 16:28:05 crc kubenswrapper[4744]: I1003 16:28:05.238717 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e" exitCode=0 Oct 03 16:28:05 crc kubenswrapper[4744]: I1003 16:28:05.238778 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e"} Oct 03 16:28:05 crc kubenswrapper[4744]: I1003 16:28:05.238853 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"ebd68629db9100ac91296997be91b152693b5453b3c435c930ae88e90650ca93"} Oct 03 16:28:05 crc kubenswrapper[4744]: I1003 16:28:05.243041 4744 generic.go:334] "Generic (PLEG): container finished" podID="f8087c3f-5904-4338-8dcf-779cf0dfa81c" containerID="7c468c596a8e7f6c898788c86efc8249d2978dd3578293514ee7f0c0ebe08487" exitCode=0 Oct 03 16:28:05 crc kubenswrapper[4744]: I1003 16:28:05.243058 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5tr5" event={"ID":"f8087c3f-5904-4338-8dcf-779cf0dfa81c","Type":"ContainerDied","Data":"7c468c596a8e7f6c898788c86efc8249d2978dd3578293514ee7f0c0ebe08487"} Oct 03 16:28:06 crc kubenswrapper[4744]: I1003 16:28:06.256841 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5tr5" event={"ID":"f8087c3f-5904-4338-8dcf-779cf0dfa81c","Type":"ContainerStarted","Data":"01ff401c15184beb4ef3d75ecb964801b1e404f8f913f7cba4313bdbfcd63481"} Oct 03 16:28:06 crc kubenswrapper[4744]: I1003 16:28:06.259891 4744 generic.go:334] "Generic (PLEG): container finished" podID="39443df9-cc38-4b33-b02e-e9136e902fc2" containerID="338cffb6c214402baef247496030860afa44ee0247f83ba4fadcab1226312d9d" exitCode=0 Oct 03 16:28:06 crc kubenswrapper[4744]: I1003 16:28:06.259958 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2m27n" event={"ID":"39443df9-cc38-4b33-b02e-e9136e902fc2","Type":"ContainerDied","Data":"338cffb6c214402baef247496030860afa44ee0247f83ba4fadcab1226312d9d"} Oct 03 16:28:06 crc kubenswrapper[4744]: I1003 16:28:06.308025 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v5tr5" podStartSLOduration=3.073756348 podStartE2EDuration="52.307995125s" podCreationTimestamp="2025-10-03 16:27:14 +0000 UTC" firstStartedPulling="2025-10-03 16:27:16.454780428 +0000 UTC m=+162.734656334" lastFinishedPulling="2025-10-03 16:28:05.689019205 +0000 UTC m=+211.968895111" observedRunningTime="2025-10-03 16:28:06.28672793 +0000 UTC m=+212.566603876" watchObservedRunningTime="2025-10-03 16:28:06.307995125 +0000 UTC m=+212.587871031" Oct 03 16:28:07 crc kubenswrapper[4744]: I1003 16:28:07.279854 4744 generic.go:334] "Generic (PLEG): container finished" podID="8be1fdf2-9a41-455e-8315-459cca9f8a09" containerID="fe5caedab13dcc6de095505d7684d6f38c2d0ef1d08060ffcfb93be8813364d2" exitCode=0 Oct 03 16:28:07 crc kubenswrapper[4744]: I1003 16:28:07.280419 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mv2tt" event={"ID":"8be1fdf2-9a41-455e-8315-459cca9f8a09","Type":"ContainerDied","Data":"fe5caedab13dcc6de095505d7684d6f38c2d0ef1d08060ffcfb93be8813364d2"} Oct 03 16:28:07 crc kubenswrapper[4744]: I1003 16:28:07.437716 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:28:07 crc kubenswrapper[4744]: I1003 16:28:07.438166 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:28:07 crc kubenswrapper[4744]: I1003 16:28:07.813100 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:28:08 crc kubenswrapper[4744]: I1003 16:28:08.288841 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2m27n" event={"ID":"39443df9-cc38-4b33-b02e-e9136e902fc2","Type":"ContainerStarted","Data":"cd30e894e09e064cd296a026fc73ac77060f6c79b3fc1c15b33f9a67e5d14e26"} Oct 03 16:28:08 crc kubenswrapper[4744]: I1003 16:28:08.332142 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:28:09 crc kubenswrapper[4744]: I1003 16:28:09.297400 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nh227" event={"ID":"18d30e6e-1a94-4d75-8b09-e01c3eb2575e","Type":"ContainerStarted","Data":"04b07a01a98c58b5cf6dd5e872176b6c07162f589c28a6eb39281b26503f8f0d"} Oct 03 16:28:09 crc kubenswrapper[4744]: I1003 16:28:09.301918 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mv2tt" event={"ID":"8be1fdf2-9a41-455e-8315-459cca9f8a09","Type":"ContainerStarted","Data":"65a4b61325f39e09ddc70e8ec9b5df55935f0ec226a6976dc2bd82ad11f11f6a"} Oct 03 16:28:09 crc kubenswrapper[4744]: I1003 16:28:09.339013 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mv2tt" podStartSLOduration=2.815522026 podStartE2EDuration="55.338983024s" podCreationTimestamp="2025-10-03 16:27:14 +0000 UTC" firstStartedPulling="2025-10-03 16:27:16.475456714 +0000 UTC m=+162.755332610" lastFinishedPulling="2025-10-03 16:28:08.998917722 +0000 UTC m=+215.278793608" observedRunningTime="2025-10-03 16:28:09.338712796 +0000 UTC m=+215.618588692" watchObservedRunningTime="2025-10-03 16:28:09.338983024 +0000 UTC m=+215.618858920" Oct 03 16:28:09 crc kubenswrapper[4744]: I1003 16:28:09.917779 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2m27n" podStartSLOduration=5.3502544610000005 podStartE2EDuration="52.917758567s" podCreationTimestamp="2025-10-03 16:27:17 +0000 UTC" firstStartedPulling="2025-10-03 16:27:19.635055824 +0000 UTC m=+165.914931720" lastFinishedPulling="2025-10-03 16:28:07.20255993 +0000 UTC m=+213.482435826" observedRunningTime="2025-10-03 16:28:09.372969459 +0000 UTC m=+215.652845365" watchObservedRunningTime="2025-10-03 16:28:09.917758567 +0000 UTC m=+216.197634463" Oct 03 16:28:10 crc kubenswrapper[4744]: I1003 16:28:10.310931 4744 generic.go:334] "Generic (PLEG): container finished" podID="18d30e6e-1a94-4d75-8b09-e01c3eb2575e" containerID="04b07a01a98c58b5cf6dd5e872176b6c07162f589c28a6eb39281b26503f8f0d" exitCode=0 Oct 03 16:28:10 crc kubenswrapper[4744]: I1003 16:28:10.311001 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nh227" event={"ID":"18d30e6e-1a94-4d75-8b09-e01c3eb2575e","Type":"ContainerDied","Data":"04b07a01a98c58b5cf6dd5e872176b6c07162f589c28a6eb39281b26503f8f0d"} Oct 03 16:28:10 crc kubenswrapper[4744]: I1003 16:28:10.313970 4744 generic.go:334] "Generic (PLEG): container finished" podID="09ab4fbf-98db-4f41-8155-19a9c7cab1ef" containerID="73caf019352d543b6182baf5082c24945a600b5243fdc7be8e3a4409a6c6f829" exitCode=0 Oct 03 16:28:10 crc kubenswrapper[4744]: I1003 16:28:10.314058 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2cdj" event={"ID":"09ab4fbf-98db-4f41-8155-19a9c7cab1ef","Type":"ContainerDied","Data":"73caf019352d543b6182baf5082c24945a600b5243fdc7be8e3a4409a6c6f829"} Oct 03 16:28:14 crc kubenswrapper[4744]: I1003 16:28:14.711852 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:28:14 crc kubenswrapper[4744]: I1003 16:28:14.712426 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:28:14 crc kubenswrapper[4744]: I1003 16:28:14.763869 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:28:14 crc kubenswrapper[4744]: I1003 16:28:14.876281 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:28:14 crc kubenswrapper[4744]: I1003 16:28:14.876370 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:28:14 crc kubenswrapper[4744]: I1003 16:28:14.916542 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:28:15 crc kubenswrapper[4744]: I1003 16:28:15.377523 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:28:15 crc kubenswrapper[4744]: I1003 16:28:15.390010 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:28:16 crc kubenswrapper[4744]: I1003 16:28:16.323180 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v5tr5"] Oct 03 16:28:17 crc kubenswrapper[4744]: I1003 16:28:17.350782 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v5tr5" podUID="f8087c3f-5904-4338-8dcf-779cf0dfa81c" containerName="registry-server" containerID="cri-o://01ff401c15184beb4ef3d75ecb964801b1e404f8f913f7cba4313bdbfcd63481" gracePeriod=2 Oct 03 16:28:17 crc kubenswrapper[4744]: I1003 16:28:17.859893 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:28:17 crc kubenswrapper[4744]: I1003 16:28:17.859932 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:28:17 crc kubenswrapper[4744]: I1003 16:28:17.898842 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:28:18 crc kubenswrapper[4744]: I1003 16:28:18.126256 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mv2tt"] Oct 03 16:28:18 crc kubenswrapper[4744]: I1003 16:28:18.126561 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mv2tt" podUID="8be1fdf2-9a41-455e-8315-459cca9f8a09" containerName="registry-server" containerID="cri-o://65a4b61325f39e09ddc70e8ec9b5df55935f0ec226a6976dc2bd82ad11f11f6a" gracePeriod=2 Oct 03 16:28:18 crc kubenswrapper[4744]: I1003 16:28:18.395139 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.365111 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mv2tt" event={"ID":"8be1fdf2-9a41-455e-8315-459cca9f8a09","Type":"ContainerDied","Data":"65a4b61325f39e09ddc70e8ec9b5df55935f0ec226a6976dc2bd82ad11f11f6a"} Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.365437 4744 generic.go:334] "Generic (PLEG): container finished" podID="8be1fdf2-9a41-455e-8315-459cca9f8a09" containerID="65a4b61325f39e09ddc70e8ec9b5df55935f0ec226a6976dc2bd82ad11f11f6a" exitCode=0 Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.370356 4744 generic.go:334] "Generic (PLEG): container finished" podID="f8087c3f-5904-4338-8dcf-779cf0dfa81c" containerID="01ff401c15184beb4ef3d75ecb964801b1e404f8f913f7cba4313bdbfcd63481" exitCode=0 Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.370412 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5tr5" event={"ID":"f8087c3f-5904-4338-8dcf-779cf0dfa81c","Type":"ContainerDied","Data":"01ff401c15184beb4ef3d75ecb964801b1e404f8f913f7cba4313bdbfcd63481"} Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.626925 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.747400 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8087c3f-5904-4338-8dcf-779cf0dfa81c-catalog-content\") pod \"f8087c3f-5904-4338-8dcf-779cf0dfa81c\" (UID: \"f8087c3f-5904-4338-8dcf-779cf0dfa81c\") " Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.747488 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8087c3f-5904-4338-8dcf-779cf0dfa81c-utilities\") pod \"f8087c3f-5904-4338-8dcf-779cf0dfa81c\" (UID: \"f8087c3f-5904-4338-8dcf-779cf0dfa81c\") " Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.747553 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vqzf\" (UniqueName: \"kubernetes.io/projected/f8087c3f-5904-4338-8dcf-779cf0dfa81c-kube-api-access-7vqzf\") pod \"f8087c3f-5904-4338-8dcf-779cf0dfa81c\" (UID: \"f8087c3f-5904-4338-8dcf-779cf0dfa81c\") " Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.748601 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8087c3f-5904-4338-8dcf-779cf0dfa81c-utilities" (OuterVolumeSpecName: "utilities") pod "f8087c3f-5904-4338-8dcf-779cf0dfa81c" (UID: "f8087c3f-5904-4338-8dcf-779cf0dfa81c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.754485 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8087c3f-5904-4338-8dcf-779cf0dfa81c-kube-api-access-7vqzf" (OuterVolumeSpecName: "kube-api-access-7vqzf") pod "f8087c3f-5904-4338-8dcf-779cf0dfa81c" (UID: "f8087c3f-5904-4338-8dcf-779cf0dfa81c"). InnerVolumeSpecName "kube-api-access-7vqzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.793841 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8087c3f-5904-4338-8dcf-779cf0dfa81c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8087c3f-5904-4338-8dcf-779cf0dfa81c" (UID: "f8087c3f-5904-4338-8dcf-779cf0dfa81c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.817694 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.852350 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vqzf\" (UniqueName: \"kubernetes.io/projected/f8087c3f-5904-4338-8dcf-779cf0dfa81c-kube-api-access-7vqzf\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.852382 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8087c3f-5904-4338-8dcf-779cf0dfa81c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.852546 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8087c3f-5904-4338-8dcf-779cf0dfa81c-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.953423 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxm8k\" (UniqueName: \"kubernetes.io/projected/8be1fdf2-9a41-455e-8315-459cca9f8a09-kube-api-access-lxm8k\") pod \"8be1fdf2-9a41-455e-8315-459cca9f8a09\" (UID: \"8be1fdf2-9a41-455e-8315-459cca9f8a09\") " Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.953481 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be1fdf2-9a41-455e-8315-459cca9f8a09-catalog-content\") pod \"8be1fdf2-9a41-455e-8315-459cca9f8a09\" (UID: \"8be1fdf2-9a41-455e-8315-459cca9f8a09\") " Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.953525 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be1fdf2-9a41-455e-8315-459cca9f8a09-utilities\") pod \"8be1fdf2-9a41-455e-8315-459cca9f8a09\" (UID: \"8be1fdf2-9a41-455e-8315-459cca9f8a09\") " Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.954967 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8be1fdf2-9a41-455e-8315-459cca9f8a09-utilities" (OuterVolumeSpecName: "utilities") pod "8be1fdf2-9a41-455e-8315-459cca9f8a09" (UID: "8be1fdf2-9a41-455e-8315-459cca9f8a09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:28:19 crc kubenswrapper[4744]: I1003 16:28:19.956487 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8be1fdf2-9a41-455e-8315-459cca9f8a09-kube-api-access-lxm8k" (OuterVolumeSpecName: "kube-api-access-lxm8k") pod "8be1fdf2-9a41-455e-8315-459cca9f8a09" (UID: "8be1fdf2-9a41-455e-8315-459cca9f8a09"). InnerVolumeSpecName "kube-api-access-lxm8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.004083 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8be1fdf2-9a41-455e-8315-459cca9f8a09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8be1fdf2-9a41-455e-8315-459cca9f8a09" (UID: "8be1fdf2-9a41-455e-8315-459cca9f8a09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.055159 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxm8k\" (UniqueName: \"kubernetes.io/projected/8be1fdf2-9a41-455e-8315-459cca9f8a09-kube-api-access-lxm8k\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.055809 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be1fdf2-9a41-455e-8315-459cca9f8a09-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.055820 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be1fdf2-9a41-455e-8315-459cca9f8a09-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.380378 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5tr5" event={"ID":"f8087c3f-5904-4338-8dcf-779cf0dfa81c","Type":"ContainerDied","Data":"9433a4d48c3463a4e9699e6eb670093d66ba96d39aff7607e5c2ee78b41f10df"} Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.380436 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5tr5" Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.380528 4744 scope.go:117] "RemoveContainer" containerID="01ff401c15184beb4ef3d75ecb964801b1e404f8f913f7cba4313bdbfcd63481" Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.383963 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mv2tt" event={"ID":"8be1fdf2-9a41-455e-8315-459cca9f8a09","Type":"ContainerDied","Data":"a6fa806a84299aa83f1b7ce09e33063ad95ff3ccab2967ff0dea127b50a813a5"} Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.384226 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mv2tt" Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.425305 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v5tr5"] Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.432701 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v5tr5"] Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.437222 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mv2tt"] Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.440512 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mv2tt"] Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.498785 4744 scope.go:117] "RemoveContainer" containerID="7c468c596a8e7f6c898788c86efc8249d2978dd3578293514ee7f0c0ebe08487" Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.529016 4744 scope.go:117] "RemoveContainer" containerID="18897b7c8387301650528ece35085546a8d97cdd2d8f99292d1a122540497e92" Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.552362 4744 scope.go:117] "RemoveContainer" containerID="65a4b61325f39e09ddc70e8ec9b5df55935f0ec226a6976dc2bd82ad11f11f6a" Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.571343 4744 scope.go:117] "RemoveContainer" containerID="fe5caedab13dcc6de095505d7684d6f38c2d0ef1d08060ffcfb93be8813364d2" Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.645428 4744 scope.go:117] "RemoveContainer" containerID="ae2f5ecd5917dd71ddd27e612b5a2df04b7d330a31e8722d293f00b97ddad8a3" Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.899998 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8be1fdf2-9a41-455e-8315-459cca9f8a09" path="/var/lib/kubelet/pods/8be1fdf2-9a41-455e-8315-459cca9f8a09/volumes" Oct 03 16:28:20 crc kubenswrapper[4744]: I1003 16:28:20.900852 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8087c3f-5904-4338-8dcf-779cf0dfa81c" path="/var/lib/kubelet/pods/f8087c3f-5904-4338-8dcf-779cf0dfa81c/volumes" Oct 03 16:28:21 crc kubenswrapper[4744]: I1003 16:28:21.122793 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2m27n"] Oct 03 16:28:21 crc kubenswrapper[4744]: I1003 16:28:21.123158 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2m27n" podUID="39443df9-cc38-4b33-b02e-e9136e902fc2" containerName="registry-server" containerID="cri-o://cd30e894e09e064cd296a026fc73ac77060f6c79b3fc1c15b33f9a67e5d14e26" gracePeriod=2 Oct 03 16:28:21 crc kubenswrapper[4744]: I1003 16:28:21.392131 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vn7c" event={"ID":"7570f2af-fefc-4726-a9ec-1c27d7cefe43","Type":"ContainerStarted","Data":"a4cd0e24e9ee116de4d722fdd64db26f86771d1e256d095b49f521f20627c223"} Oct 03 16:28:21 crc kubenswrapper[4744]: I1003 16:28:21.394913 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nh227" event={"ID":"18d30e6e-1a94-4d75-8b09-e01c3eb2575e","Type":"ContainerStarted","Data":"bc89ab140c4372136786b4cb4730f19aa5a651bef248269b4e0e28d22bb8cf8e"} Oct 03 16:28:21 crc kubenswrapper[4744]: I1003 16:28:21.397282 4744 generic.go:334] "Generic (PLEG): container finished" podID="1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" containerID="7de818ac7fd2acd97e30bc9031888c6a1a2f9589452579f0809428a588621f07" exitCode=0 Oct 03 16:28:21 crc kubenswrapper[4744]: I1003 16:28:21.397341 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lsw7x" event={"ID":"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056","Type":"ContainerDied","Data":"7de818ac7fd2acd97e30bc9031888c6a1a2f9589452579f0809428a588621f07"} Oct 03 16:28:21 crc kubenswrapper[4744]: I1003 16:28:21.400660 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2cdj" event={"ID":"09ab4fbf-98db-4f41-8155-19a9c7cab1ef","Type":"ContainerStarted","Data":"f9cba9c0b7e7b435aaa77997bb83980c1d2e630e074815cd2617bcdf56c4130f"} Oct 03 16:28:21 crc kubenswrapper[4744]: I1003 16:28:21.452884 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nh227" podStartSLOduration=3.393301947 podStartE2EDuration="1m7.452858671s" podCreationTimestamp="2025-10-03 16:27:14 +0000 UTC" firstStartedPulling="2025-10-03 16:27:16.440344964 +0000 UTC m=+162.720220860" lastFinishedPulling="2025-10-03 16:28:20.499901678 +0000 UTC m=+226.779777584" observedRunningTime="2025-10-03 16:28:21.451127653 +0000 UTC m=+227.731003569" watchObservedRunningTime="2025-10-03 16:28:21.452858671 +0000 UTC m=+227.732734567" Oct 03 16:28:21 crc kubenswrapper[4744]: I1003 16:28:21.478727 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z2cdj" podStartSLOduration=3.645865733 podStartE2EDuration="1m6.478706113s" podCreationTimestamp="2025-10-03 16:27:15 +0000 UTC" firstStartedPulling="2025-10-03 16:27:17.503051613 +0000 UTC m=+163.782927509" lastFinishedPulling="2025-10-03 16:28:20.335891993 +0000 UTC m=+226.615767889" observedRunningTime="2025-10-03 16:28:21.473467188 +0000 UTC m=+227.753343084" watchObservedRunningTime="2025-10-03 16:28:21.478706113 +0000 UTC m=+227.758582009" Oct 03 16:28:22 crc kubenswrapper[4744]: I1003 16:28:22.421112 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lsw7x" event={"ID":"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056","Type":"ContainerStarted","Data":"829f90777832c33f1e261834c81e654ee3a5f81677f6e3a15f1f1bf6fd99859c"} Oct 03 16:28:22 crc kubenswrapper[4744]: I1003 16:28:22.428957 4744 generic.go:334] "Generic (PLEG): container finished" podID="39443df9-cc38-4b33-b02e-e9136e902fc2" containerID="cd30e894e09e064cd296a026fc73ac77060f6c79b3fc1c15b33f9a67e5d14e26" exitCode=0 Oct 03 16:28:22 crc kubenswrapper[4744]: I1003 16:28:22.429062 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2m27n" event={"ID":"39443df9-cc38-4b33-b02e-e9136e902fc2","Type":"ContainerDied","Data":"cd30e894e09e064cd296a026fc73ac77060f6c79b3fc1c15b33f9a67e5d14e26"} Oct 03 16:28:22 crc kubenswrapper[4744]: I1003 16:28:22.440961 4744 generic.go:334] "Generic (PLEG): container finished" podID="7570f2af-fefc-4726-a9ec-1c27d7cefe43" containerID="a4cd0e24e9ee116de4d722fdd64db26f86771d1e256d095b49f521f20627c223" exitCode=0 Oct 03 16:28:22 crc kubenswrapper[4744]: I1003 16:28:22.441018 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vn7c" event={"ID":"7570f2af-fefc-4726-a9ec-1c27d7cefe43","Type":"ContainerDied","Data":"a4cd0e24e9ee116de4d722fdd64db26f86771d1e256d095b49f521f20627c223"} Oct 03 16:28:22 crc kubenswrapper[4744]: I1003 16:28:22.455428 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lsw7x" podStartSLOduration=2.14128662 podStartE2EDuration="1m6.4554069s" podCreationTimestamp="2025-10-03 16:27:16 +0000 UTC" firstStartedPulling="2025-10-03 16:27:17.522281199 +0000 UTC m=+163.802157095" lastFinishedPulling="2025-10-03 16:28:21.836401469 +0000 UTC m=+228.116277375" observedRunningTime="2025-10-03 16:28:22.452725566 +0000 UTC m=+228.732601482" watchObservedRunningTime="2025-10-03 16:28:22.4554069 +0000 UTC m=+228.735282796" Oct 03 16:28:22 crc kubenswrapper[4744]: I1003 16:28:22.629986 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:28:22 crc kubenswrapper[4744]: I1003 16:28:22.791581 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39443df9-cc38-4b33-b02e-e9136e902fc2-catalog-content\") pod \"39443df9-cc38-4b33-b02e-e9136e902fc2\" (UID: \"39443df9-cc38-4b33-b02e-e9136e902fc2\") " Oct 03 16:28:22 crc kubenswrapper[4744]: I1003 16:28:22.791678 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39443df9-cc38-4b33-b02e-e9136e902fc2-utilities\") pod \"39443df9-cc38-4b33-b02e-e9136e902fc2\" (UID: \"39443df9-cc38-4b33-b02e-e9136e902fc2\") " Oct 03 16:28:22 crc kubenswrapper[4744]: I1003 16:28:22.791784 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2vrm\" (UniqueName: \"kubernetes.io/projected/39443df9-cc38-4b33-b02e-e9136e902fc2-kube-api-access-v2vrm\") pod \"39443df9-cc38-4b33-b02e-e9136e902fc2\" (UID: \"39443df9-cc38-4b33-b02e-e9136e902fc2\") " Oct 03 16:28:22 crc kubenswrapper[4744]: I1003 16:28:22.792572 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39443df9-cc38-4b33-b02e-e9136e902fc2-utilities" (OuterVolumeSpecName: "utilities") pod "39443df9-cc38-4b33-b02e-e9136e902fc2" (UID: "39443df9-cc38-4b33-b02e-e9136e902fc2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:28:22 crc kubenswrapper[4744]: I1003 16:28:22.812373 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39443df9-cc38-4b33-b02e-e9136e902fc2-kube-api-access-v2vrm" (OuterVolumeSpecName: "kube-api-access-v2vrm") pod "39443df9-cc38-4b33-b02e-e9136e902fc2" (UID: "39443df9-cc38-4b33-b02e-e9136e902fc2"). InnerVolumeSpecName "kube-api-access-v2vrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:28:22 crc kubenswrapper[4744]: I1003 16:28:22.893865 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39443df9-cc38-4b33-b02e-e9136e902fc2-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:22 crc kubenswrapper[4744]: I1003 16:28:22.897139 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39443df9-cc38-4b33-b02e-e9136e902fc2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "39443df9-cc38-4b33-b02e-e9136e902fc2" (UID: "39443df9-cc38-4b33-b02e-e9136e902fc2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:28:22 crc kubenswrapper[4744]: I1003 16:28:22.909033 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2vrm\" (UniqueName: \"kubernetes.io/projected/39443df9-cc38-4b33-b02e-e9136e902fc2-kube-api-access-v2vrm\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:23 crc kubenswrapper[4744]: I1003 16:28:23.010607 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39443df9-cc38-4b33-b02e-e9136e902fc2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:23 crc kubenswrapper[4744]: I1003 16:28:23.451513 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vn7c" event={"ID":"7570f2af-fefc-4726-a9ec-1c27d7cefe43","Type":"ContainerStarted","Data":"770eabf0f5455c87f8a8e8634a8eb25de9cdff0631ba5f813c613118e374cb9a"} Oct 03 16:28:23 crc kubenswrapper[4744]: I1003 16:28:23.454967 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2m27n" event={"ID":"39443df9-cc38-4b33-b02e-e9136e902fc2","Type":"ContainerDied","Data":"9e928f51ec30ac785424065feb0ff395379a63c6204db8ce5305dd09bbf3f051"} Oct 03 16:28:23 crc kubenswrapper[4744]: I1003 16:28:23.455045 4744 scope.go:117] "RemoveContainer" containerID="cd30e894e09e064cd296a026fc73ac77060f6c79b3fc1c15b33f9a67e5d14e26" Oct 03 16:28:23 crc kubenswrapper[4744]: I1003 16:28:23.455134 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2m27n" Oct 03 16:28:23 crc kubenswrapper[4744]: I1003 16:28:23.471690 4744 scope.go:117] "RemoveContainer" containerID="338cffb6c214402baef247496030860afa44ee0247f83ba4fadcab1226312d9d" Oct 03 16:28:23 crc kubenswrapper[4744]: I1003 16:28:23.482285 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7vn7c" podStartSLOduration=2.822064955 podStartE2EDuration="1m10.482266448s" podCreationTimestamp="2025-10-03 16:27:13 +0000 UTC" firstStartedPulling="2025-10-03 16:27:15.394825494 +0000 UTC m=+161.674701390" lastFinishedPulling="2025-10-03 16:28:23.055026987 +0000 UTC m=+229.334902883" observedRunningTime="2025-10-03 16:28:23.479592095 +0000 UTC m=+229.759467981" watchObservedRunningTime="2025-10-03 16:28:23.482266448 +0000 UTC m=+229.762142344" Oct 03 16:28:23 crc kubenswrapper[4744]: I1003 16:28:23.498550 4744 scope.go:117] "RemoveContainer" containerID="5f0fe5f98666cf3af5a95ac6585c9546bb04e5e02f564de4332efccb264d6dfc" Oct 03 16:28:23 crc kubenswrapper[4744]: I1003 16:28:23.500188 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2m27n"] Oct 03 16:28:23 crc kubenswrapper[4744]: I1003 16:28:23.505014 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2m27n"] Oct 03 16:28:24 crc kubenswrapper[4744]: I1003 16:28:24.235059 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:28:24 crc kubenswrapper[4744]: I1003 16:28:24.235196 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:28:24 crc kubenswrapper[4744]: I1003 16:28:24.737143 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:28:24 crc kubenswrapper[4744]: I1003 16:28:24.737587 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:28:24 crc kubenswrapper[4744]: I1003 16:28:24.803006 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:28:24 crc kubenswrapper[4744]: I1003 16:28:24.899590 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39443df9-cc38-4b33-b02e-e9136e902fc2" path="/var/lib/kubelet/pods/39443df9-cc38-4b33-b02e-e9136e902fc2/volumes" Oct 03 16:28:25 crc kubenswrapper[4744]: I1003 16:28:25.276900 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-7vn7c" podUID="7570f2af-fefc-4726-a9ec-1c27d7cefe43" containerName="registry-server" probeResult="failure" output=< Oct 03 16:28:25 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 03 16:28:25 crc kubenswrapper[4744]: > Oct 03 16:28:26 crc kubenswrapper[4744]: I1003 16:28:26.275312 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:28:26 crc kubenswrapper[4744]: I1003 16:28:26.275389 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:28:26 crc kubenswrapper[4744]: I1003 16:28:26.322179 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:28:26 crc kubenswrapper[4744]: I1003 16:28:26.513998 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:28:26 crc kubenswrapper[4744]: I1003 16:28:26.648007 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:28:26 crc kubenswrapper[4744]: I1003 16:28:26.648081 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:28:26 crc kubenswrapper[4744]: I1003 16:28:26.696527 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:28:26 crc kubenswrapper[4744]: I1003 16:28:26.807119 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6jxqm"] Oct 03 16:28:27 crc kubenswrapper[4744]: I1003 16:28:27.535223 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:28:28 crc kubenswrapper[4744]: I1003 16:28:28.923383 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lsw7x"] Oct 03 16:28:29 crc kubenswrapper[4744]: I1003 16:28:29.499304 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lsw7x" podUID="1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" containerName="registry-server" containerID="cri-o://829f90777832c33f1e261834c81e654ee3a5f81677f6e3a15f1f1bf6fd99859c" gracePeriod=2 Oct 03 16:28:29 crc kubenswrapper[4744]: I1003 16:28:29.906581 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.018940 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-utilities\") pod \"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056\" (UID: \"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056\") " Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.019168 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-catalog-content\") pod \"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056\" (UID: \"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056\") " Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.019243 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sssg\" (UniqueName: \"kubernetes.io/projected/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-kube-api-access-8sssg\") pod \"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056\" (UID: \"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056\") " Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.021151 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-utilities" (OuterVolumeSpecName: "utilities") pod "1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" (UID: "1bb5b864-3bb2-4e32-9e0a-63f54c1f8056"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.026349 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-kube-api-access-8sssg" (OuterVolumeSpecName: "kube-api-access-8sssg") pod "1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" (UID: "1bb5b864-3bb2-4e32-9e0a-63f54c1f8056"). InnerVolumeSpecName "kube-api-access-8sssg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.042209 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" (UID: "1bb5b864-3bb2-4e32-9e0a-63f54c1f8056"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.121384 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.121432 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sssg\" (UniqueName: \"kubernetes.io/projected/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-kube-api-access-8sssg\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.121449 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.507240 4744 generic.go:334] "Generic (PLEG): container finished" podID="1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" containerID="829f90777832c33f1e261834c81e654ee3a5f81677f6e3a15f1f1bf6fd99859c" exitCode=0 Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.507298 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lsw7x" event={"ID":"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056","Type":"ContainerDied","Data":"829f90777832c33f1e261834c81e654ee3a5f81677f6e3a15f1f1bf6fd99859c"} Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.507333 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lsw7x" event={"ID":"1bb5b864-3bb2-4e32-9e0a-63f54c1f8056","Type":"ContainerDied","Data":"83037ff1f46bca709bd33a9e26da6e55599d2a4ac8c323370330e393ebd506ea"} Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.507353 4744 scope.go:117] "RemoveContainer" containerID="829f90777832c33f1e261834c81e654ee3a5f81677f6e3a15f1f1bf6fd99859c" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.507536 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lsw7x" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.541407 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lsw7x"] Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.545850 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lsw7x"] Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.547109 4744 scope.go:117] "RemoveContainer" containerID="7de818ac7fd2acd97e30bc9031888c6a1a2f9589452579f0809428a588621f07" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.562219 4744 scope.go:117] "RemoveContainer" containerID="e0bd0e4f690ef101b1e80ded5d26752b2a88b7bd81742bc2f065322d5f5df47b" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.581077 4744 scope.go:117] "RemoveContainer" containerID="829f90777832c33f1e261834c81e654ee3a5f81677f6e3a15f1f1bf6fd99859c" Oct 03 16:28:30 crc kubenswrapper[4744]: E1003 16:28:30.581610 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"829f90777832c33f1e261834c81e654ee3a5f81677f6e3a15f1f1bf6fd99859c\": container with ID starting with 829f90777832c33f1e261834c81e654ee3a5f81677f6e3a15f1f1bf6fd99859c not found: ID does not exist" containerID="829f90777832c33f1e261834c81e654ee3a5f81677f6e3a15f1f1bf6fd99859c" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.581668 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"829f90777832c33f1e261834c81e654ee3a5f81677f6e3a15f1f1bf6fd99859c"} err="failed to get container status \"829f90777832c33f1e261834c81e654ee3a5f81677f6e3a15f1f1bf6fd99859c\": rpc error: code = NotFound desc = could not find container \"829f90777832c33f1e261834c81e654ee3a5f81677f6e3a15f1f1bf6fd99859c\": container with ID starting with 829f90777832c33f1e261834c81e654ee3a5f81677f6e3a15f1f1bf6fd99859c not found: ID does not exist" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.581703 4744 scope.go:117] "RemoveContainer" containerID="7de818ac7fd2acd97e30bc9031888c6a1a2f9589452579f0809428a588621f07" Oct 03 16:28:30 crc kubenswrapper[4744]: E1003 16:28:30.582236 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7de818ac7fd2acd97e30bc9031888c6a1a2f9589452579f0809428a588621f07\": container with ID starting with 7de818ac7fd2acd97e30bc9031888c6a1a2f9589452579f0809428a588621f07 not found: ID does not exist" containerID="7de818ac7fd2acd97e30bc9031888c6a1a2f9589452579f0809428a588621f07" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.582278 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7de818ac7fd2acd97e30bc9031888c6a1a2f9589452579f0809428a588621f07"} err="failed to get container status \"7de818ac7fd2acd97e30bc9031888c6a1a2f9589452579f0809428a588621f07\": rpc error: code = NotFound desc = could not find container \"7de818ac7fd2acd97e30bc9031888c6a1a2f9589452579f0809428a588621f07\": container with ID starting with 7de818ac7fd2acd97e30bc9031888c6a1a2f9589452579f0809428a588621f07 not found: ID does not exist" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.582315 4744 scope.go:117] "RemoveContainer" containerID="e0bd0e4f690ef101b1e80ded5d26752b2a88b7bd81742bc2f065322d5f5df47b" Oct 03 16:28:30 crc kubenswrapper[4744]: E1003 16:28:30.582638 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0bd0e4f690ef101b1e80ded5d26752b2a88b7bd81742bc2f065322d5f5df47b\": container with ID starting with e0bd0e4f690ef101b1e80ded5d26752b2a88b7bd81742bc2f065322d5f5df47b not found: ID does not exist" containerID="e0bd0e4f690ef101b1e80ded5d26752b2a88b7bd81742bc2f065322d5f5df47b" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.582670 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0bd0e4f690ef101b1e80ded5d26752b2a88b7bd81742bc2f065322d5f5df47b"} err="failed to get container status \"e0bd0e4f690ef101b1e80ded5d26752b2a88b7bd81742bc2f065322d5f5df47b\": rpc error: code = NotFound desc = could not find container \"e0bd0e4f690ef101b1e80ded5d26752b2a88b7bd81742bc2f065322d5f5df47b\": container with ID starting with e0bd0e4f690ef101b1e80ded5d26752b2a88b7bd81742bc2f065322d5f5df47b not found: ID does not exist" Oct 03 16:28:30 crc kubenswrapper[4744]: I1003 16:28:30.899256 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" path="/var/lib/kubelet/pods/1bb5b864-3bb2-4e32-9e0a-63f54c1f8056/volumes" Oct 03 16:28:34 crc kubenswrapper[4744]: I1003 16:28:34.280277 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:28:34 crc kubenswrapper[4744]: I1003 16:28:34.322204 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:28:34 crc kubenswrapper[4744]: I1003 16:28:34.780563 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:28:51 crc kubenswrapper[4744]: I1003 16:28:51.837559 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" podUID="4f39c88b-8d2a-4a19-921e-14d8d172ad20" containerName="oauth-openshift" containerID="cri-o://9daf5dec3e1dfc007bd7c8e2369b05cb67283ba709124f784ec8c34066e25897" gracePeriod=15 Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.222665 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.256234 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5969b76fdc-nxggg"] Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.256713 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee70474c-bce0-4289-af10-7e1ae1393b5a" containerName="pruner" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.256816 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee70474c-bce0-4289-af10-7e1ae1393b5a" containerName="pruner" Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.256888 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be1fdf2-9a41-455e-8315-459cca9f8a09" containerName="registry-server" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.256958 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be1fdf2-9a41-455e-8315-459cca9f8a09" containerName="registry-server" Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.257033 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f39c88b-8d2a-4a19-921e-14d8d172ad20" containerName="oauth-openshift" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.257093 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f39c88b-8d2a-4a19-921e-14d8d172ad20" containerName="oauth-openshift" Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.257155 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be1fdf2-9a41-455e-8315-459cca9f8a09" containerName="extract-utilities" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.257215 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be1fdf2-9a41-455e-8315-459cca9f8a09" containerName="extract-utilities" Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.257274 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8087c3f-5904-4338-8dcf-779cf0dfa81c" containerName="registry-server" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.257330 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8087c3f-5904-4338-8dcf-779cf0dfa81c" containerName="registry-server" Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.257437 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39443df9-cc38-4b33-b02e-e9136e902fc2" containerName="extract-utilities" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.257513 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="39443df9-cc38-4b33-b02e-e9136e902fc2" containerName="extract-utilities" Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.257588 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39443df9-cc38-4b33-b02e-e9136e902fc2" containerName="extract-content" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.257648 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="39443df9-cc38-4b33-b02e-e9136e902fc2" containerName="extract-content" Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.257702 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" containerName="extract-content" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.257759 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" containerName="extract-content" Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.257813 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8087c3f-5904-4338-8dcf-779cf0dfa81c" containerName="extract-utilities" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.257871 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8087c3f-5904-4338-8dcf-779cf0dfa81c" containerName="extract-utilities" Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.257924 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" containerName="registry-server" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.257980 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" containerName="registry-server" Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.258048 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be1fdf2-9a41-455e-8315-459cca9f8a09" containerName="extract-content" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.258188 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be1fdf2-9a41-455e-8315-459cca9f8a09" containerName="extract-content" Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.258326 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39443df9-cc38-4b33-b02e-e9136e902fc2" containerName="registry-server" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.258425 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="39443df9-cc38-4b33-b02e-e9136e902fc2" containerName="registry-server" Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.258566 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8087c3f-5904-4338-8dcf-779cf0dfa81c" containerName="extract-content" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.258655 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8087c3f-5904-4338-8dcf-779cf0dfa81c" containerName="extract-content" Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.259020 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" containerName="extract-utilities" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.259106 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" containerName="extract-utilities" Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.259213 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9de21d2-7004-4b21-bf31-ef1c280aa940" containerName="pruner" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.259287 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9de21d2-7004-4b21-bf31-ef1c280aa940" containerName="pruner" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.259474 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bb5b864-3bb2-4e32-9e0a-63f54c1f8056" containerName="registry-server" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.259688 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee70474c-bce0-4289-af10-7e1ae1393b5a" containerName="pruner" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.259804 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="39443df9-cc38-4b33-b02e-e9136e902fc2" containerName="registry-server" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.259900 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8087c3f-5904-4338-8dcf-779cf0dfa81c" containerName="registry-server" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.260001 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9de21d2-7004-4b21-bf31-ef1c280aa940" containerName="pruner" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.260083 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8be1fdf2-9a41-455e-8315-459cca9f8a09" containerName="registry-server" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.260163 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f39c88b-8d2a-4a19-921e-14d8d172ad20" containerName="oauth-openshift" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.260849 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.271316 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5969b76fdc-nxggg"] Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357378 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n6qv\" (UniqueName: \"kubernetes.io/projected/4f39c88b-8d2a-4a19-921e-14d8d172ad20-kube-api-access-2n6qv\") pod \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357435 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4f39c88b-8d2a-4a19-921e-14d8d172ad20-audit-dir\") pod \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357505 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-cliconfig\") pod \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357526 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-serving-cert\") pod \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357588 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-session\") pod \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357613 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-provider-selection\") pod \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357635 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-trusted-ca-bundle\") pod \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357670 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-router-certs\") pod \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357700 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-ocp-branding-template\") pod \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357722 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-service-ca\") pod \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357742 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-login\") pod \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357764 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-audit-policies\") pod \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357791 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-idp-0-file-data\") pod \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357815 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-error\") pod \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\" (UID: \"4f39c88b-8d2a-4a19-921e-14d8d172ad20\") " Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357966 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.357997 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.358015 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-user-template-login\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.358032 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/af357b5e-87bf-4a60-a95c-406fcf60a49b-audit-policies\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.358054 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.358080 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/af357b5e-87bf-4a60-a95c-406fcf60a49b-audit-dir\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.358101 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjfjj\" (UniqueName: \"kubernetes.io/projected/af357b5e-87bf-4a60-a95c-406fcf60a49b-kube-api-access-sjfjj\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.358143 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.358165 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-user-template-error\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.358188 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-session\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.358207 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.358230 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.358247 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-service-ca\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.358266 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-router-certs\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.359700 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f39c88b-8d2a-4a19-921e-14d8d172ad20-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "4f39c88b-8d2a-4a19-921e-14d8d172ad20" (UID: "4f39c88b-8d2a-4a19-921e-14d8d172ad20"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.359985 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "4f39c88b-8d2a-4a19-921e-14d8d172ad20" (UID: "4f39c88b-8d2a-4a19-921e-14d8d172ad20"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.360559 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "4f39c88b-8d2a-4a19-921e-14d8d172ad20" (UID: "4f39c88b-8d2a-4a19-921e-14d8d172ad20"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.360892 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "4f39c88b-8d2a-4a19-921e-14d8d172ad20" (UID: "4f39c88b-8d2a-4a19-921e-14d8d172ad20"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.361486 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "4f39c88b-8d2a-4a19-921e-14d8d172ad20" (UID: "4f39c88b-8d2a-4a19-921e-14d8d172ad20"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.366166 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "4f39c88b-8d2a-4a19-921e-14d8d172ad20" (UID: "4f39c88b-8d2a-4a19-921e-14d8d172ad20"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.366253 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f39c88b-8d2a-4a19-921e-14d8d172ad20-kube-api-access-2n6qv" (OuterVolumeSpecName: "kube-api-access-2n6qv") pod "4f39c88b-8d2a-4a19-921e-14d8d172ad20" (UID: "4f39c88b-8d2a-4a19-921e-14d8d172ad20"). InnerVolumeSpecName "kube-api-access-2n6qv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.366895 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "4f39c88b-8d2a-4a19-921e-14d8d172ad20" (UID: "4f39c88b-8d2a-4a19-921e-14d8d172ad20"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.367118 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "4f39c88b-8d2a-4a19-921e-14d8d172ad20" (UID: "4f39c88b-8d2a-4a19-921e-14d8d172ad20"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.367936 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "4f39c88b-8d2a-4a19-921e-14d8d172ad20" (UID: "4f39c88b-8d2a-4a19-921e-14d8d172ad20"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.368328 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "4f39c88b-8d2a-4a19-921e-14d8d172ad20" (UID: "4f39c88b-8d2a-4a19-921e-14d8d172ad20"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.370335 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "4f39c88b-8d2a-4a19-921e-14d8d172ad20" (UID: "4f39c88b-8d2a-4a19-921e-14d8d172ad20"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.372981 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "4f39c88b-8d2a-4a19-921e-14d8d172ad20" (UID: "4f39c88b-8d2a-4a19-921e-14d8d172ad20"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.373350 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "4f39c88b-8d2a-4a19-921e-14d8d172ad20" (UID: "4f39c88b-8d2a-4a19-921e-14d8d172ad20"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.459865 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.459945 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-user-template-login\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.460212 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.460251 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/af357b5e-87bf-4a60-a95c-406fcf60a49b-audit-policies\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.460884 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461186 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461253 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/af357b5e-87bf-4a60-a95c-406fcf60a49b-audit-dir\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461284 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjfjj\" (UniqueName: \"kubernetes.io/projected/af357b5e-87bf-4a60-a95c-406fcf60a49b-kube-api-access-sjfjj\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461323 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/af357b5e-87bf-4a60-a95c-406fcf60a49b-audit-policies\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461335 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461417 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/af357b5e-87bf-4a60-a95c-406fcf60a49b-audit-dir\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461462 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-user-template-error\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461437 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461560 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-session\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461611 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461679 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461717 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-service-ca\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461743 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-router-certs\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461850 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n6qv\" (UniqueName: \"kubernetes.io/projected/4f39c88b-8d2a-4a19-921e-14d8d172ad20-kube-api-access-2n6qv\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461877 4744 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4f39c88b-8d2a-4a19-921e-14d8d172ad20-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461889 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461902 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461913 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461927 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461938 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461951 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461962 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461974 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461986 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.461997 4744 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4f39c88b-8d2a-4a19-921e-14d8d172ad20-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.462008 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.462022 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4f39c88b-8d2a-4a19-921e-14d8d172ad20-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.462621 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-service-ca\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.466275 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-user-template-error\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.466524 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.466603 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.466648 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-router-certs\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.466770 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-user-template-login\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.467021 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-session\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.467264 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.469046 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/af357b5e-87bf-4a60-a95c-406fcf60a49b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.483740 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjfjj\" (UniqueName: \"kubernetes.io/projected/af357b5e-87bf-4a60-a95c-406fcf60a49b-kube-api-access-sjfjj\") pod \"oauth-openshift-5969b76fdc-nxggg\" (UID: \"af357b5e-87bf-4a60-a95c-406fcf60a49b\") " pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.596879 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.652049 4744 generic.go:334] "Generic (PLEG): container finished" podID="4f39c88b-8d2a-4a19-921e-14d8d172ad20" containerID="9daf5dec3e1dfc007bd7c8e2369b05cb67283ba709124f784ec8c34066e25897" exitCode=0 Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.652119 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" event={"ID":"4f39c88b-8d2a-4a19-921e-14d8d172ad20","Type":"ContainerDied","Data":"9daf5dec3e1dfc007bd7c8e2369b05cb67283ba709124f784ec8c34066e25897"} Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.652170 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" event={"ID":"4f39c88b-8d2a-4a19-921e-14d8d172ad20","Type":"ContainerDied","Data":"1b565c39e26d076105a185b95e565027488a1d2fc456f55e024f38e20c52151b"} Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.652195 4744 scope.go:117] "RemoveContainer" containerID="9daf5dec3e1dfc007bd7c8e2369b05cb67283ba709124f784ec8c34066e25897" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.652190 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6jxqm" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.714888 4744 scope.go:117] "RemoveContainer" containerID="9daf5dec3e1dfc007bd7c8e2369b05cb67283ba709124f784ec8c34066e25897" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.715187 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6jxqm"] Oct 03 16:28:52 crc kubenswrapper[4744]: E1003 16:28:52.715802 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9daf5dec3e1dfc007bd7c8e2369b05cb67283ba709124f784ec8c34066e25897\": container with ID starting with 9daf5dec3e1dfc007bd7c8e2369b05cb67283ba709124f784ec8c34066e25897 not found: ID does not exist" containerID="9daf5dec3e1dfc007bd7c8e2369b05cb67283ba709124f784ec8c34066e25897" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.715832 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9daf5dec3e1dfc007bd7c8e2369b05cb67283ba709124f784ec8c34066e25897"} err="failed to get container status \"9daf5dec3e1dfc007bd7c8e2369b05cb67283ba709124f784ec8c34066e25897\": rpc error: code = NotFound desc = could not find container \"9daf5dec3e1dfc007bd7c8e2369b05cb67283ba709124f784ec8c34066e25897\": container with ID starting with 9daf5dec3e1dfc007bd7c8e2369b05cb67283ba709124f784ec8c34066e25897 not found: ID does not exist" Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.720097 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6jxqm"] Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.833707 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5969b76fdc-nxggg"] Oct 03 16:28:52 crc kubenswrapper[4744]: I1003 16:28:52.900249 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f39c88b-8d2a-4a19-921e-14d8d172ad20" path="/var/lib/kubelet/pods/4f39c88b-8d2a-4a19-921e-14d8d172ad20/volumes" Oct 03 16:28:53 crc kubenswrapper[4744]: I1003 16:28:53.663734 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" event={"ID":"af357b5e-87bf-4a60-a95c-406fcf60a49b","Type":"ContainerStarted","Data":"b2cdb83ef533a0cdda6a77a994457cdba10116cb260e27c464b1fb7478412355"} Oct 03 16:28:53 crc kubenswrapper[4744]: I1003 16:28:53.664195 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" event={"ID":"af357b5e-87bf-4a60-a95c-406fcf60a49b","Type":"ContainerStarted","Data":"7f3f6668aed555b89586f4684a89031cb6be8392025c00f9ff3ed37052d22483"} Oct 03 16:28:53 crc kubenswrapper[4744]: I1003 16:28:53.664316 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:53 crc kubenswrapper[4744]: I1003 16:28:53.678190 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" Oct 03 16:28:53 crc kubenswrapper[4744]: I1003 16:28:53.699169 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5969b76fdc-nxggg" podStartSLOduration=27.699140675 podStartE2EDuration="27.699140675s" podCreationTimestamp="2025-10-03 16:28:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:28:53.694994151 +0000 UTC m=+259.974870057" watchObservedRunningTime="2025-10-03 16:28:53.699140675 +0000 UTC m=+259.979016581" Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.413211 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nh227"] Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.416259 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nh227" podUID="18d30e6e-1a94-4d75-8b09-e01c3eb2575e" containerName="registry-server" containerID="cri-o://bc89ab140c4372136786b4cb4730f19aa5a651bef248269b4e0e28d22bb8cf8e" gracePeriod=30 Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.428974 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7vn7c"] Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.429376 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7vn7c" podUID="7570f2af-fefc-4726-a9ec-1c27d7cefe43" containerName="registry-server" containerID="cri-o://770eabf0f5455c87f8a8e8634a8eb25de9cdff0631ba5f813c613118e374cb9a" gracePeriod=30 Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.439746 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9hm6h"] Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.440106 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" podUID="cfcd176c-f3fd-419f-92ad-f7fdbcab6270" containerName="marketplace-operator" containerID="cri-o://41ae90d5fc75256882279e151f4940c81ba81084375f4ba5a2c808f95f4a4a89" gracePeriod=30 Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.446429 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2cdj"] Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.446838 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z2cdj" podUID="09ab4fbf-98db-4f41-8155-19a9c7cab1ef" containerName="registry-server" containerID="cri-o://f9cba9c0b7e7b435aaa77997bb83980c1d2e630e074815cd2617bcdf56c4130f" gracePeriod=30 Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.458726 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7jvbs"] Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.459109 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7jvbs" podUID="7b5a6267-63e7-44e7-a891-2ab9c9b8b29b" containerName="registry-server" containerID="cri-o://9c31a04e5a9f04d92476c63fc866a57202ae33eb8a3a996e498f4294cf616bfb" gracePeriod=30 Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.462760 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xr8s7"] Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.463764 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.472067 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xr8s7"] Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.610653 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bddf7be1-c68d-42ab-853e-153710114dd4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xr8s7\" (UID: \"bddf7be1-c68d-42ab-853e-153710114dd4\") " pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.610709 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bddf7be1-c68d-42ab-853e-153710114dd4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xr8s7\" (UID: \"bddf7be1-c68d-42ab-853e-153710114dd4\") " pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.610766 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kht4\" (UniqueName: \"kubernetes.io/projected/bddf7be1-c68d-42ab-853e-153710114dd4-kube-api-access-7kht4\") pod \"marketplace-operator-79b997595-xr8s7\" (UID: \"bddf7be1-c68d-42ab-853e-153710114dd4\") " pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.712428 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bddf7be1-c68d-42ab-853e-153710114dd4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xr8s7\" (UID: \"bddf7be1-c68d-42ab-853e-153710114dd4\") " pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.712473 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bddf7be1-c68d-42ab-853e-153710114dd4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xr8s7\" (UID: \"bddf7be1-c68d-42ab-853e-153710114dd4\") " pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.712570 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kht4\" (UniqueName: \"kubernetes.io/projected/bddf7be1-c68d-42ab-853e-153710114dd4-kube-api-access-7kht4\") pod \"marketplace-operator-79b997595-xr8s7\" (UID: \"bddf7be1-c68d-42ab-853e-153710114dd4\") " pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.714817 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bddf7be1-c68d-42ab-853e-153710114dd4-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xr8s7\" (UID: \"bddf7be1-c68d-42ab-853e-153710114dd4\") " pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.725807 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bddf7be1-c68d-42ab-853e-153710114dd4-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xr8s7\" (UID: \"bddf7be1-c68d-42ab-853e-153710114dd4\") " pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.730012 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kht4\" (UniqueName: \"kubernetes.io/projected/bddf7be1-c68d-42ab-853e-153710114dd4-kube-api-access-7kht4\") pod \"marketplace-operator-79b997595-xr8s7\" (UID: \"bddf7be1-c68d-42ab-853e-153710114dd4\") " pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" Oct 03 16:29:14 crc kubenswrapper[4744]: E1003 16:29:14.738006 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bc89ab140c4372136786b4cb4730f19aa5a651bef248269b4e0e28d22bb8cf8e is running failed: container process not found" containerID="bc89ab140c4372136786b4cb4730f19aa5a651bef248269b4e0e28d22bb8cf8e" cmd=["grpc_health_probe","-addr=:50051"] Oct 03 16:29:14 crc kubenswrapper[4744]: E1003 16:29:14.738424 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bc89ab140c4372136786b4cb4730f19aa5a651bef248269b4e0e28d22bb8cf8e is running failed: container process not found" containerID="bc89ab140c4372136786b4cb4730f19aa5a651bef248269b4e0e28d22bb8cf8e" cmd=["grpc_health_probe","-addr=:50051"] Oct 03 16:29:14 crc kubenswrapper[4744]: E1003 16:29:14.738722 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bc89ab140c4372136786b4cb4730f19aa5a651bef248269b4e0e28d22bb8cf8e is running failed: container process not found" containerID="bc89ab140c4372136786b4cb4730f19aa5a651bef248269b4e0e28d22bb8cf8e" cmd=["grpc_health_probe","-addr=:50051"] Oct 03 16:29:14 crc kubenswrapper[4744]: E1003 16:29:14.738762 4744 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of bc89ab140c4372136786b4cb4730f19aa5a651bef248269b4e0e28d22bb8cf8e is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-nh227" podUID="18d30e6e-1a94-4d75-8b09-e01c3eb2575e" containerName="registry-server" Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.789009 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.814583 4744 generic.go:334] "Generic (PLEG): container finished" podID="18d30e6e-1a94-4d75-8b09-e01c3eb2575e" containerID="bc89ab140c4372136786b4cb4730f19aa5a651bef248269b4e0e28d22bb8cf8e" exitCode=0 Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.814698 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nh227" event={"ID":"18d30e6e-1a94-4d75-8b09-e01c3eb2575e","Type":"ContainerDied","Data":"bc89ab140c4372136786b4cb4730f19aa5a651bef248269b4e0e28d22bb8cf8e"} Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.829064 4744 generic.go:334] "Generic (PLEG): container finished" podID="09ab4fbf-98db-4f41-8155-19a9c7cab1ef" containerID="f9cba9c0b7e7b435aaa77997bb83980c1d2e630e074815cd2617bcdf56c4130f" exitCode=0 Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.829156 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2cdj" event={"ID":"09ab4fbf-98db-4f41-8155-19a9c7cab1ef","Type":"ContainerDied","Data":"f9cba9c0b7e7b435aaa77997bb83980c1d2e630e074815cd2617bcdf56c4130f"} Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.831961 4744 generic.go:334] "Generic (PLEG): container finished" podID="7b5a6267-63e7-44e7-a891-2ab9c9b8b29b" containerID="9c31a04e5a9f04d92476c63fc866a57202ae33eb8a3a996e498f4294cf616bfb" exitCode=0 Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.832022 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jvbs" event={"ID":"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b","Type":"ContainerDied","Data":"9c31a04e5a9f04d92476c63fc866a57202ae33eb8a3a996e498f4294cf616bfb"} Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.850843 4744 generic.go:334] "Generic (PLEG): container finished" podID="cfcd176c-f3fd-419f-92ad-f7fdbcab6270" containerID="41ae90d5fc75256882279e151f4940c81ba81084375f4ba5a2c808f95f4a4a89" exitCode=0 Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.850928 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" event={"ID":"cfcd176c-f3fd-419f-92ad-f7fdbcab6270","Type":"ContainerDied","Data":"41ae90d5fc75256882279e151f4940c81ba81084375f4ba5a2c808f95f4a4a89"} Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.863180 4744 generic.go:334] "Generic (PLEG): container finished" podID="7570f2af-fefc-4726-a9ec-1c27d7cefe43" containerID="770eabf0f5455c87f8a8e8634a8eb25de9cdff0631ba5f813c613118e374cb9a" exitCode=0 Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.863240 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vn7c" event={"ID":"7570f2af-fefc-4726-a9ec-1c27d7cefe43","Type":"ContainerDied","Data":"770eabf0f5455c87f8a8e8634a8eb25de9cdff0631ba5f813c613118e374cb9a"} Oct 03 16:29:14 crc kubenswrapper[4744]: I1003 16:29:14.903669 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.016406 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c49hm\" (UniqueName: \"kubernetes.io/projected/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-kube-api-access-c49hm\") pod \"cfcd176c-f3fd-419f-92ad-f7fdbcab6270\" (UID: \"cfcd176c-f3fd-419f-92ad-f7fdbcab6270\") " Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.016533 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-marketplace-trusted-ca\") pod \"cfcd176c-f3fd-419f-92ad-f7fdbcab6270\" (UID: \"cfcd176c-f3fd-419f-92ad-f7fdbcab6270\") " Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.016559 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-marketplace-operator-metrics\") pod \"cfcd176c-f3fd-419f-92ad-f7fdbcab6270\" (UID: \"cfcd176c-f3fd-419f-92ad-f7fdbcab6270\") " Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.018734 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "cfcd176c-f3fd-419f-92ad-f7fdbcab6270" (UID: "cfcd176c-f3fd-419f-92ad-f7fdbcab6270"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.029843 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-kube-api-access-c49hm" (OuterVolumeSpecName: "kube-api-access-c49hm") pod "cfcd176c-f3fd-419f-92ad-f7fdbcab6270" (UID: "cfcd176c-f3fd-419f-92ad-f7fdbcab6270"). InnerVolumeSpecName "kube-api-access-c49hm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.030085 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "cfcd176c-f3fd-419f-92ad-f7fdbcab6270" (UID: "cfcd176c-f3fd-419f-92ad-f7fdbcab6270"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.076235 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.089543 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.099469 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.119311 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c49hm\" (UniqueName: \"kubernetes.io/projected/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-kube-api-access-c49hm\") on node \"crc\" DevicePath \"\"" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.119346 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.119355 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cfcd176c-f3fd-419f-92ad-f7fdbcab6270-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.210323 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.220068 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh2lw\" (UniqueName: \"kubernetes.io/projected/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-kube-api-access-lh2lw\") pod \"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b\" (UID: \"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b\") " Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.220148 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7570f2af-fefc-4726-a9ec-1c27d7cefe43-utilities\") pod \"7570f2af-fefc-4726-a9ec-1c27d7cefe43\" (UID: \"7570f2af-fefc-4726-a9ec-1c27d7cefe43\") " Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.220169 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-catalog-content\") pod \"18d30e6e-1a94-4d75-8b09-e01c3eb2575e\" (UID: \"18d30e6e-1a94-4d75-8b09-e01c3eb2575e\") " Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.220198 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7570f2af-fefc-4726-a9ec-1c27d7cefe43-catalog-content\") pod \"7570f2af-fefc-4726-a9ec-1c27d7cefe43\" (UID: \"7570f2af-fefc-4726-a9ec-1c27d7cefe43\") " Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.220229 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk9qh\" (UniqueName: \"kubernetes.io/projected/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-kube-api-access-hk9qh\") pod \"18d30e6e-1a94-4d75-8b09-e01c3eb2575e\" (UID: \"18d30e6e-1a94-4d75-8b09-e01c3eb2575e\") " Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.220278 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-utilities\") pod \"18d30e6e-1a94-4d75-8b09-e01c3eb2575e\" (UID: \"18d30e6e-1a94-4d75-8b09-e01c3eb2575e\") " Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.220323 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-catalog-content\") pod \"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b\" (UID: \"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b\") " Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.220345 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpl6l\" (UniqueName: \"kubernetes.io/projected/7570f2af-fefc-4726-a9ec-1c27d7cefe43-kube-api-access-kpl6l\") pod \"7570f2af-fefc-4726-a9ec-1c27d7cefe43\" (UID: \"7570f2af-fefc-4726-a9ec-1c27d7cefe43\") " Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.220375 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-utilities\") pod \"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b\" (UID: \"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b\") " Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.220967 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7570f2af-fefc-4726-a9ec-1c27d7cefe43-utilities" (OuterVolumeSpecName: "utilities") pod "7570f2af-fefc-4726-a9ec-1c27d7cefe43" (UID: "7570f2af-fefc-4726-a9ec-1c27d7cefe43"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.221313 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-utilities" (OuterVolumeSpecName: "utilities") pod "7b5a6267-63e7-44e7-a891-2ab9c9b8b29b" (UID: "7b5a6267-63e7-44e7-a891-2ab9c9b8b29b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.222603 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-utilities" (OuterVolumeSpecName: "utilities") pod "18d30e6e-1a94-4d75-8b09-e01c3eb2575e" (UID: "18d30e6e-1a94-4d75-8b09-e01c3eb2575e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.227718 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-kube-api-access-hk9qh" (OuterVolumeSpecName: "kube-api-access-hk9qh") pod "18d30e6e-1a94-4d75-8b09-e01c3eb2575e" (UID: "18d30e6e-1a94-4d75-8b09-e01c3eb2575e"). InnerVolumeSpecName "kube-api-access-hk9qh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.235796 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-kube-api-access-lh2lw" (OuterVolumeSpecName: "kube-api-access-lh2lw") pod "7b5a6267-63e7-44e7-a891-2ab9c9b8b29b" (UID: "7b5a6267-63e7-44e7-a891-2ab9c9b8b29b"). InnerVolumeSpecName "kube-api-access-lh2lw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.249705 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7570f2af-fefc-4726-a9ec-1c27d7cefe43-kube-api-access-kpl6l" (OuterVolumeSpecName: "kube-api-access-kpl6l") pod "7570f2af-fefc-4726-a9ec-1c27d7cefe43" (UID: "7570f2af-fefc-4726-a9ec-1c27d7cefe43"). InnerVolumeSpecName "kube-api-access-kpl6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.293465 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7570f2af-fefc-4726-a9ec-1c27d7cefe43-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7570f2af-fefc-4726-a9ec-1c27d7cefe43" (UID: "7570f2af-fefc-4726-a9ec-1c27d7cefe43"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.301668 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18d30e6e-1a94-4d75-8b09-e01c3eb2575e" (UID: "18d30e6e-1a94-4d75-8b09-e01c3eb2575e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.321800 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npsjm\" (UniqueName: \"kubernetes.io/projected/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-kube-api-access-npsjm\") pod \"09ab4fbf-98db-4f41-8155-19a9c7cab1ef\" (UID: \"09ab4fbf-98db-4f41-8155-19a9c7cab1ef\") " Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.321859 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-catalog-content\") pod \"09ab4fbf-98db-4f41-8155-19a9c7cab1ef\" (UID: \"09ab4fbf-98db-4f41-8155-19a9c7cab1ef\") " Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.321904 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-utilities\") pod \"09ab4fbf-98db-4f41-8155-19a9c7cab1ef\" (UID: \"09ab4fbf-98db-4f41-8155-19a9c7cab1ef\") " Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.322109 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh2lw\" (UniqueName: \"kubernetes.io/projected/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-kube-api-access-lh2lw\") on node \"crc\" DevicePath \"\"" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.322123 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7570f2af-fefc-4726-a9ec-1c27d7cefe43-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.322135 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.322155 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7570f2af-fefc-4726-a9ec-1c27d7cefe43-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.322169 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk9qh\" (UniqueName: \"kubernetes.io/projected/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-kube-api-access-hk9qh\") on node \"crc\" DevicePath \"\"" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.322181 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18d30e6e-1a94-4d75-8b09-e01c3eb2575e-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.322192 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpl6l\" (UniqueName: \"kubernetes.io/projected/7570f2af-fefc-4726-a9ec-1c27d7cefe43-kube-api-access-kpl6l\") on node \"crc\" DevicePath \"\"" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.322202 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.323079 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-utilities" (OuterVolumeSpecName: "utilities") pod "09ab4fbf-98db-4f41-8155-19a9c7cab1ef" (UID: "09ab4fbf-98db-4f41-8155-19a9c7cab1ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.325639 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-kube-api-access-npsjm" (OuterVolumeSpecName: "kube-api-access-npsjm") pod "09ab4fbf-98db-4f41-8155-19a9c7cab1ef" (UID: "09ab4fbf-98db-4f41-8155-19a9c7cab1ef"). InnerVolumeSpecName "kube-api-access-npsjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.334881 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09ab4fbf-98db-4f41-8155-19a9c7cab1ef" (UID: "09ab4fbf-98db-4f41-8155-19a9c7cab1ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.341047 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b5a6267-63e7-44e7-a891-2ab9c9b8b29b" (UID: "7b5a6267-63e7-44e7-a891-2ab9c9b8b29b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.423694 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.423732 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npsjm\" (UniqueName: \"kubernetes.io/projected/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-kube-api-access-npsjm\") on node \"crc\" DevicePath \"\"" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.423753 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.423767 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09ab4fbf-98db-4f41-8155-19a9c7cab1ef-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.446762 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xr8s7"] Oct 03 16:29:15 crc kubenswrapper[4744]: W1003 16:29:15.453732 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbddf7be1_c68d_42ab_853e_153710114dd4.slice/crio-bd55c5869aab941be873c45f570743cfa0f8559e0bcb60606f3f23fb03de79f5 WatchSource:0}: Error finding container bd55c5869aab941be873c45f570743cfa0f8559e0bcb60606f3f23fb03de79f5: Status 404 returned error can't find the container with id bd55c5869aab941be873c45f570743cfa0f8559e0bcb60606f3f23fb03de79f5 Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.869941 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" event={"ID":"bddf7be1-c68d-42ab-853e-153710114dd4","Type":"ContainerStarted","Data":"db9dcdf34f9e46e191321aca3201f477c11bbe6fc5efcc577e5f0bfdae9ed05f"} Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.870022 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" event={"ID":"bddf7be1-c68d-42ab-853e-153710114dd4","Type":"ContainerStarted","Data":"bd55c5869aab941be873c45f570743cfa0f8559e0bcb60606f3f23fb03de79f5"} Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.870330 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.872462 4744 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xr8s7 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.872537 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" podUID="bddf7be1-c68d-42ab-853e-153710114dd4" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.872861 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vn7c" event={"ID":"7570f2af-fefc-4726-a9ec-1c27d7cefe43","Type":"ContainerDied","Data":"4ba2119cb5dea1c393e5224d33213d0f30614a543f0afd687299d1b428b33c15"} Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.872895 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7vn7c" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.872994 4744 scope.go:117] "RemoveContainer" containerID="770eabf0f5455c87f8a8e8634a8eb25de9cdff0631ba5f813c613118e374cb9a" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.875511 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nh227" event={"ID":"18d30e6e-1a94-4d75-8b09-e01c3eb2575e","Type":"ContainerDied","Data":"3cdfd94c609ea4e3994bbd3f1dd85b6f20c5e21b6c3fd175e82bbb5b6598fb8e"} Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.875563 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nh227" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.881898 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2cdj" event={"ID":"09ab4fbf-98db-4f41-8155-19a9c7cab1ef","Type":"ContainerDied","Data":"6441d2c33c5ef0c47bb49f40f445e60acbd0c87fcc1b76926c190471376e156e"} Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.882053 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2cdj" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.885277 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jvbs" event={"ID":"7b5a6267-63e7-44e7-a891-2ab9c9b8b29b","Type":"ContainerDied","Data":"82eb980126a1a1f312e22a9a12c87a2520b46d9fb5dee253218bfe2940f4b268"} Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.885411 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jvbs" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.893025 4744 scope.go:117] "RemoveContainer" containerID="a4cd0e24e9ee116de4d722fdd64db26f86771d1e256d095b49f521f20627c223" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.895179 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" podStartSLOduration=1.895158414 podStartE2EDuration="1.895158414s" podCreationTimestamp="2025-10-03 16:29:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:29:15.893468176 +0000 UTC m=+282.173344072" watchObservedRunningTime="2025-10-03 16:29:15.895158414 +0000 UTC m=+282.175034310" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.896031 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" event={"ID":"cfcd176c-f3fd-419f-92ad-f7fdbcab6270","Type":"ContainerDied","Data":"286a60c0444c4cf67a80c02e6137b130e8c68a28048b05284a094712b3b7cfe1"} Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.896112 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9hm6h" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.932959 4744 scope.go:117] "RemoveContainer" containerID="3a3c46c528827bacdb2e2e655308cc5eab31fb776a07a5e776e175d2424695d3" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.961826 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nh227"] Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.966664 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nh227"] Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.969546 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7jvbs"] Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.978952 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7jvbs"] Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.980521 4744 scope.go:117] "RemoveContainer" containerID="bc89ab140c4372136786b4cb4730f19aa5a651bef248269b4e0e28d22bb8cf8e" Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.985990 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7vn7c"] Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.992222 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7vn7c"] Oct 03 16:29:15 crc kubenswrapper[4744]: I1003 16:29:15.999786 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2cdj"] Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.001733 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2cdj"] Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.003887 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9hm6h"] Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.006002 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9hm6h"] Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.008614 4744 scope.go:117] "RemoveContainer" containerID="04b07a01a98c58b5cf6dd5e872176b6c07162f589c28a6eb39281b26503f8f0d" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.026969 4744 scope.go:117] "RemoveContainer" containerID="b0d9eab7d606f87bc36239bcb43357a78e098dfa00b4952f6884c6e47e5d9edb" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.044104 4744 scope.go:117] "RemoveContainer" containerID="f9cba9c0b7e7b435aaa77997bb83980c1d2e630e074815cd2617bcdf56c4130f" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.066166 4744 scope.go:117] "RemoveContainer" containerID="73caf019352d543b6182baf5082c24945a600b5243fdc7be8e3a4409a6c6f829" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.083158 4744 scope.go:117] "RemoveContainer" containerID="afa91b7a2c29f18be0e6cf3b37b460db5808f4ddf381dbf79fbf8b9840deb89e" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.099167 4744 scope.go:117] "RemoveContainer" containerID="9c31a04e5a9f04d92476c63fc866a57202ae33eb8a3a996e498f4294cf616bfb" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.111895 4744 scope.go:117] "RemoveContainer" containerID="09822b10139977124afca9bc3e46b3c5d9af0eaa3d2509c4f60914084e0c7e2d" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.124997 4744 scope.go:117] "RemoveContainer" containerID="035308c2a9ef1e069c6af81885bcad9b0bed87952ee1d25c2fd87e08ab65ecca" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.140551 4744 scope.go:117] "RemoveContainer" containerID="41ae90d5fc75256882279e151f4940c81ba81084375f4ba5a2c808f95f4a4a89" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644196 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wmqbq"] Oct 03 16:29:16 crc kubenswrapper[4744]: E1003 16:29:16.644505 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18d30e6e-1a94-4d75-8b09-e01c3eb2575e" containerName="extract-utilities" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644522 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="18d30e6e-1a94-4d75-8b09-e01c3eb2575e" containerName="extract-utilities" Oct 03 16:29:16 crc kubenswrapper[4744]: E1003 16:29:16.644534 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b5a6267-63e7-44e7-a891-2ab9c9b8b29b" containerName="extract-utilities" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644546 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b5a6267-63e7-44e7-a891-2ab9c9b8b29b" containerName="extract-utilities" Oct 03 16:29:16 crc kubenswrapper[4744]: E1003 16:29:16.644558 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18d30e6e-1a94-4d75-8b09-e01c3eb2575e" containerName="registry-server" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644566 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="18d30e6e-1a94-4d75-8b09-e01c3eb2575e" containerName="registry-server" Oct 03 16:29:16 crc kubenswrapper[4744]: E1003 16:29:16.644576 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfcd176c-f3fd-419f-92ad-f7fdbcab6270" containerName="marketplace-operator" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644583 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfcd176c-f3fd-419f-92ad-f7fdbcab6270" containerName="marketplace-operator" Oct 03 16:29:16 crc kubenswrapper[4744]: E1003 16:29:16.644594 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09ab4fbf-98db-4f41-8155-19a9c7cab1ef" containerName="extract-utilities" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644601 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="09ab4fbf-98db-4f41-8155-19a9c7cab1ef" containerName="extract-utilities" Oct 03 16:29:16 crc kubenswrapper[4744]: E1003 16:29:16.644614 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09ab4fbf-98db-4f41-8155-19a9c7cab1ef" containerName="registry-server" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644620 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="09ab4fbf-98db-4f41-8155-19a9c7cab1ef" containerName="registry-server" Oct 03 16:29:16 crc kubenswrapper[4744]: E1003 16:29:16.644629 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18d30e6e-1a94-4d75-8b09-e01c3eb2575e" containerName="extract-content" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644635 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="18d30e6e-1a94-4d75-8b09-e01c3eb2575e" containerName="extract-content" Oct 03 16:29:16 crc kubenswrapper[4744]: E1003 16:29:16.644644 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7570f2af-fefc-4726-a9ec-1c27d7cefe43" containerName="extract-content" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644650 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7570f2af-fefc-4726-a9ec-1c27d7cefe43" containerName="extract-content" Oct 03 16:29:16 crc kubenswrapper[4744]: E1003 16:29:16.644660 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b5a6267-63e7-44e7-a891-2ab9c9b8b29b" containerName="registry-server" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644668 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b5a6267-63e7-44e7-a891-2ab9c9b8b29b" containerName="registry-server" Oct 03 16:29:16 crc kubenswrapper[4744]: E1003 16:29:16.644683 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7570f2af-fefc-4726-a9ec-1c27d7cefe43" containerName="registry-server" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644690 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7570f2af-fefc-4726-a9ec-1c27d7cefe43" containerName="registry-server" Oct 03 16:29:16 crc kubenswrapper[4744]: E1003 16:29:16.644703 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b5a6267-63e7-44e7-a891-2ab9c9b8b29b" containerName="extract-content" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644710 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b5a6267-63e7-44e7-a891-2ab9c9b8b29b" containerName="extract-content" Oct 03 16:29:16 crc kubenswrapper[4744]: E1003 16:29:16.644721 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09ab4fbf-98db-4f41-8155-19a9c7cab1ef" containerName="extract-content" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644728 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="09ab4fbf-98db-4f41-8155-19a9c7cab1ef" containerName="extract-content" Oct 03 16:29:16 crc kubenswrapper[4744]: E1003 16:29:16.644743 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7570f2af-fefc-4726-a9ec-1c27d7cefe43" containerName="extract-utilities" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644750 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7570f2af-fefc-4726-a9ec-1c27d7cefe43" containerName="extract-utilities" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644891 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfcd176c-f3fd-419f-92ad-f7fdbcab6270" containerName="marketplace-operator" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644901 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7570f2af-fefc-4726-a9ec-1c27d7cefe43" containerName="registry-server" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644911 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="18d30e6e-1a94-4d75-8b09-e01c3eb2575e" containerName="registry-server" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644921 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="09ab4fbf-98db-4f41-8155-19a9c7cab1ef" containerName="registry-server" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.644929 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b5a6267-63e7-44e7-a891-2ab9c9b8b29b" containerName="registry-server" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.645892 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wmqbq" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.648792 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.660937 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmqbq"] Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.837619 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l8gwc"] Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.839655 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l8gwc" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.844088 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86654f3a-97fa-456a-b651-93f8ec174725-catalog-content\") pod \"redhat-marketplace-wmqbq\" (UID: \"86654f3a-97fa-456a-b651-93f8ec174725\") " pod="openshift-marketplace/redhat-marketplace-wmqbq" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.844170 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86654f3a-97fa-456a-b651-93f8ec174725-utilities\") pod \"redhat-marketplace-wmqbq\" (UID: \"86654f3a-97fa-456a-b651-93f8ec174725\") " pod="openshift-marketplace/redhat-marketplace-wmqbq" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.844248 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.844259 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmz82\" (UniqueName: \"kubernetes.io/projected/86654f3a-97fa-456a-b651-93f8ec174725-kube-api-access-kmz82\") pod \"redhat-marketplace-wmqbq\" (UID: \"86654f3a-97fa-456a-b651-93f8ec174725\") " pod="openshift-marketplace/redhat-marketplace-wmqbq" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.848934 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l8gwc"] Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.902636 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ab4fbf-98db-4f41-8155-19a9c7cab1ef" path="/var/lib/kubelet/pods/09ab4fbf-98db-4f41-8155-19a9c7cab1ef/volumes" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.903480 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18d30e6e-1a94-4d75-8b09-e01c3eb2575e" path="/var/lib/kubelet/pods/18d30e6e-1a94-4d75-8b09-e01c3eb2575e/volumes" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.904110 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7570f2af-fefc-4726-a9ec-1c27d7cefe43" path="/var/lib/kubelet/pods/7570f2af-fefc-4726-a9ec-1c27d7cefe43/volumes" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.905149 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b5a6267-63e7-44e7-a891-2ab9c9b8b29b" path="/var/lib/kubelet/pods/7b5a6267-63e7-44e7-a891-2ab9c9b8b29b/volumes" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.905809 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfcd176c-f3fd-419f-92ad-f7fdbcab6270" path="/var/lib/kubelet/pods/cfcd176c-f3fd-419f-92ad-f7fdbcab6270/volumes" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.913478 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xr8s7" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.946234 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fhpg\" (UniqueName: \"kubernetes.io/projected/83f32f6a-6664-4e8f-8676-cefcc31b6e6c-kube-api-access-4fhpg\") pod \"redhat-operators-l8gwc\" (UID: \"83f32f6a-6664-4e8f-8676-cefcc31b6e6c\") " pod="openshift-marketplace/redhat-operators-l8gwc" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.946389 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmz82\" (UniqueName: \"kubernetes.io/projected/86654f3a-97fa-456a-b651-93f8ec174725-kube-api-access-kmz82\") pod \"redhat-marketplace-wmqbq\" (UID: \"86654f3a-97fa-456a-b651-93f8ec174725\") " pod="openshift-marketplace/redhat-marketplace-wmqbq" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.946440 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83f32f6a-6664-4e8f-8676-cefcc31b6e6c-utilities\") pod \"redhat-operators-l8gwc\" (UID: \"83f32f6a-6664-4e8f-8676-cefcc31b6e6c\") " pod="openshift-marketplace/redhat-operators-l8gwc" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.946519 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83f32f6a-6664-4e8f-8676-cefcc31b6e6c-catalog-content\") pod \"redhat-operators-l8gwc\" (UID: \"83f32f6a-6664-4e8f-8676-cefcc31b6e6c\") " pod="openshift-marketplace/redhat-operators-l8gwc" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.946570 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86654f3a-97fa-456a-b651-93f8ec174725-catalog-content\") pod \"redhat-marketplace-wmqbq\" (UID: \"86654f3a-97fa-456a-b651-93f8ec174725\") " pod="openshift-marketplace/redhat-marketplace-wmqbq" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.946596 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86654f3a-97fa-456a-b651-93f8ec174725-utilities\") pod \"redhat-marketplace-wmqbq\" (UID: \"86654f3a-97fa-456a-b651-93f8ec174725\") " pod="openshift-marketplace/redhat-marketplace-wmqbq" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.947204 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86654f3a-97fa-456a-b651-93f8ec174725-utilities\") pod \"redhat-marketplace-wmqbq\" (UID: \"86654f3a-97fa-456a-b651-93f8ec174725\") " pod="openshift-marketplace/redhat-marketplace-wmqbq" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.947246 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86654f3a-97fa-456a-b651-93f8ec174725-catalog-content\") pod \"redhat-marketplace-wmqbq\" (UID: \"86654f3a-97fa-456a-b651-93f8ec174725\") " pod="openshift-marketplace/redhat-marketplace-wmqbq" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.968037 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmz82\" (UniqueName: \"kubernetes.io/projected/86654f3a-97fa-456a-b651-93f8ec174725-kube-api-access-kmz82\") pod \"redhat-marketplace-wmqbq\" (UID: \"86654f3a-97fa-456a-b651-93f8ec174725\") " pod="openshift-marketplace/redhat-marketplace-wmqbq" Oct 03 16:29:16 crc kubenswrapper[4744]: I1003 16:29:16.981066 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wmqbq" Oct 03 16:29:17 crc kubenswrapper[4744]: I1003 16:29:17.047397 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83f32f6a-6664-4e8f-8676-cefcc31b6e6c-catalog-content\") pod \"redhat-operators-l8gwc\" (UID: \"83f32f6a-6664-4e8f-8676-cefcc31b6e6c\") " pod="openshift-marketplace/redhat-operators-l8gwc" Oct 03 16:29:17 crc kubenswrapper[4744]: I1003 16:29:17.047897 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fhpg\" (UniqueName: \"kubernetes.io/projected/83f32f6a-6664-4e8f-8676-cefcc31b6e6c-kube-api-access-4fhpg\") pod \"redhat-operators-l8gwc\" (UID: \"83f32f6a-6664-4e8f-8676-cefcc31b6e6c\") " pod="openshift-marketplace/redhat-operators-l8gwc" Oct 03 16:29:17 crc kubenswrapper[4744]: I1003 16:29:17.048008 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83f32f6a-6664-4e8f-8676-cefcc31b6e6c-utilities\") pod \"redhat-operators-l8gwc\" (UID: \"83f32f6a-6664-4e8f-8676-cefcc31b6e6c\") " pod="openshift-marketplace/redhat-operators-l8gwc" Oct 03 16:29:17 crc kubenswrapper[4744]: I1003 16:29:17.049216 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83f32f6a-6664-4e8f-8676-cefcc31b6e6c-utilities\") pod \"redhat-operators-l8gwc\" (UID: \"83f32f6a-6664-4e8f-8676-cefcc31b6e6c\") " pod="openshift-marketplace/redhat-operators-l8gwc" Oct 03 16:29:17 crc kubenswrapper[4744]: I1003 16:29:17.049970 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83f32f6a-6664-4e8f-8676-cefcc31b6e6c-catalog-content\") pod \"redhat-operators-l8gwc\" (UID: \"83f32f6a-6664-4e8f-8676-cefcc31b6e6c\") " pod="openshift-marketplace/redhat-operators-l8gwc" Oct 03 16:29:17 crc kubenswrapper[4744]: I1003 16:29:17.084228 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fhpg\" (UniqueName: \"kubernetes.io/projected/83f32f6a-6664-4e8f-8676-cefcc31b6e6c-kube-api-access-4fhpg\") pod \"redhat-operators-l8gwc\" (UID: \"83f32f6a-6664-4e8f-8676-cefcc31b6e6c\") " pod="openshift-marketplace/redhat-operators-l8gwc" Oct 03 16:29:17 crc kubenswrapper[4744]: I1003 16:29:17.198911 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l8gwc" Oct 03 16:29:17 crc kubenswrapper[4744]: I1003 16:29:17.402759 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wmqbq"] Oct 03 16:29:17 crc kubenswrapper[4744]: I1003 16:29:17.602010 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l8gwc"] Oct 03 16:29:17 crc kubenswrapper[4744]: W1003 16:29:17.607555 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83f32f6a_6664_4e8f_8676_cefcc31b6e6c.slice/crio-459e8867a9d5a86f3819c0fadc5fda36b7b04429b0591e69cb6f5b02d8fbfc21 WatchSource:0}: Error finding container 459e8867a9d5a86f3819c0fadc5fda36b7b04429b0591e69cb6f5b02d8fbfc21: Status 404 returned error can't find the container with id 459e8867a9d5a86f3819c0fadc5fda36b7b04429b0591e69cb6f5b02d8fbfc21 Oct 03 16:29:17 crc kubenswrapper[4744]: I1003 16:29:17.917738 4744 generic.go:334] "Generic (PLEG): container finished" podID="86654f3a-97fa-456a-b651-93f8ec174725" containerID="5c39a27b5344b87b5bafa039584bffec052c9741b42e507c27964a4f5d92a5a6" exitCode=0 Oct 03 16:29:17 crc kubenswrapper[4744]: I1003 16:29:17.918206 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmqbq" event={"ID":"86654f3a-97fa-456a-b651-93f8ec174725","Type":"ContainerDied","Data":"5c39a27b5344b87b5bafa039584bffec052c9741b42e507c27964a4f5d92a5a6"} Oct 03 16:29:17 crc kubenswrapper[4744]: I1003 16:29:17.918233 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmqbq" event={"ID":"86654f3a-97fa-456a-b651-93f8ec174725","Type":"ContainerStarted","Data":"f4cff87dc68de79c5203bcf7dd1515385da660f2927caf29b67c898b8a7c3081"} Oct 03 16:29:17 crc kubenswrapper[4744]: I1003 16:29:17.923403 4744 generic.go:334] "Generic (PLEG): container finished" podID="83f32f6a-6664-4e8f-8676-cefcc31b6e6c" containerID="aec8b14d3b5920a6a042bc921775c85533ff386744c93945ec457025a41b5042" exitCode=0 Oct 03 16:29:17 crc kubenswrapper[4744]: I1003 16:29:17.923580 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8gwc" event={"ID":"83f32f6a-6664-4e8f-8676-cefcc31b6e6c","Type":"ContainerDied","Data":"aec8b14d3b5920a6a042bc921775c85533ff386744c93945ec457025a41b5042"} Oct 03 16:29:17 crc kubenswrapper[4744]: I1003 16:29:17.923612 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8gwc" event={"ID":"83f32f6a-6664-4e8f-8676-cefcc31b6e6c","Type":"ContainerStarted","Data":"459e8867a9d5a86f3819c0fadc5fda36b7b04429b0591e69cb6f5b02d8fbfc21"} Oct 03 16:29:18 crc kubenswrapper[4744]: I1003 16:29:18.935446 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmqbq" event={"ID":"86654f3a-97fa-456a-b651-93f8ec174725","Type":"ContainerStarted","Data":"f737804e0197af9cec599f83786c41c63713176fe846adefaccab77a16c0d840"} Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.039631 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vcfbk"] Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.040941 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vcfbk" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.048346 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.050832 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vcfbk"] Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.181540 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6fb66df-a4db-417b-90fc-c779777883be-catalog-content\") pod \"certified-operators-vcfbk\" (UID: \"f6fb66df-a4db-417b-90fc-c779777883be\") " pod="openshift-marketplace/certified-operators-vcfbk" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.181633 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6fb66df-a4db-417b-90fc-c779777883be-utilities\") pod \"certified-operators-vcfbk\" (UID: \"f6fb66df-a4db-417b-90fc-c779777883be\") " pod="openshift-marketplace/certified-operators-vcfbk" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.181688 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2gjs\" (UniqueName: \"kubernetes.io/projected/f6fb66df-a4db-417b-90fc-c779777883be-kube-api-access-j2gjs\") pod \"certified-operators-vcfbk\" (UID: \"f6fb66df-a4db-417b-90fc-c779777883be\") " pod="openshift-marketplace/certified-operators-vcfbk" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.235112 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7v9wq"] Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.236466 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7v9wq" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.239222 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.244978 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7v9wq"] Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.282978 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6fb66df-a4db-417b-90fc-c779777883be-catalog-content\") pod \"certified-operators-vcfbk\" (UID: \"f6fb66df-a4db-417b-90fc-c779777883be\") " pod="openshift-marketplace/certified-operators-vcfbk" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.283051 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2-catalog-content\") pod \"community-operators-7v9wq\" (UID: \"6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2\") " pod="openshift-marketplace/community-operators-7v9wq" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.283083 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6fb66df-a4db-417b-90fc-c779777883be-utilities\") pod \"certified-operators-vcfbk\" (UID: \"f6fb66df-a4db-417b-90fc-c779777883be\") " pod="openshift-marketplace/certified-operators-vcfbk" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.283125 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2gjs\" (UniqueName: \"kubernetes.io/projected/f6fb66df-a4db-417b-90fc-c779777883be-kube-api-access-j2gjs\") pod \"certified-operators-vcfbk\" (UID: \"f6fb66df-a4db-417b-90fc-c779777883be\") " pod="openshift-marketplace/certified-operators-vcfbk" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.283155 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2-utilities\") pod \"community-operators-7v9wq\" (UID: \"6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2\") " pod="openshift-marketplace/community-operators-7v9wq" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.283183 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klkqz\" (UniqueName: \"kubernetes.io/projected/6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2-kube-api-access-klkqz\") pod \"community-operators-7v9wq\" (UID: \"6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2\") " pod="openshift-marketplace/community-operators-7v9wq" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.283573 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6fb66df-a4db-417b-90fc-c779777883be-catalog-content\") pod \"certified-operators-vcfbk\" (UID: \"f6fb66df-a4db-417b-90fc-c779777883be\") " pod="openshift-marketplace/certified-operators-vcfbk" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.292071 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6fb66df-a4db-417b-90fc-c779777883be-utilities\") pod \"certified-operators-vcfbk\" (UID: \"f6fb66df-a4db-417b-90fc-c779777883be\") " pod="openshift-marketplace/certified-operators-vcfbk" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.314152 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2gjs\" (UniqueName: \"kubernetes.io/projected/f6fb66df-a4db-417b-90fc-c779777883be-kube-api-access-j2gjs\") pod \"certified-operators-vcfbk\" (UID: \"f6fb66df-a4db-417b-90fc-c779777883be\") " pod="openshift-marketplace/certified-operators-vcfbk" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.380130 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vcfbk" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.383952 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2-catalog-content\") pod \"community-operators-7v9wq\" (UID: \"6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2\") " pod="openshift-marketplace/community-operators-7v9wq" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.384009 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2-utilities\") pod \"community-operators-7v9wq\" (UID: \"6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2\") " pod="openshift-marketplace/community-operators-7v9wq" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.384031 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klkqz\" (UniqueName: \"kubernetes.io/projected/6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2-kube-api-access-klkqz\") pod \"community-operators-7v9wq\" (UID: \"6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2\") " pod="openshift-marketplace/community-operators-7v9wq" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.384769 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2-catalog-content\") pod \"community-operators-7v9wq\" (UID: \"6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2\") " pod="openshift-marketplace/community-operators-7v9wq" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.385148 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2-utilities\") pod \"community-operators-7v9wq\" (UID: \"6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2\") " pod="openshift-marketplace/community-operators-7v9wq" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.403385 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klkqz\" (UniqueName: \"kubernetes.io/projected/6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2-kube-api-access-klkqz\") pod \"community-operators-7v9wq\" (UID: \"6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2\") " pod="openshift-marketplace/community-operators-7v9wq" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.557845 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7v9wq" Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.793545 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vcfbk"] Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.943736 4744 generic.go:334] "Generic (PLEG): container finished" podID="86654f3a-97fa-456a-b651-93f8ec174725" containerID="f737804e0197af9cec599f83786c41c63713176fe846adefaccab77a16c0d840" exitCode=0 Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.944278 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmqbq" event={"ID":"86654f3a-97fa-456a-b651-93f8ec174725","Type":"ContainerDied","Data":"f737804e0197af9cec599f83786c41c63713176fe846adefaccab77a16c0d840"} Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.950014 4744 generic.go:334] "Generic (PLEG): container finished" podID="83f32f6a-6664-4e8f-8676-cefcc31b6e6c" containerID="c22de03ac8c963cc300d233de1c8334cfdc65d1699252c9acfbf415608aba387" exitCode=0 Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.950168 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8gwc" event={"ID":"83f32f6a-6664-4e8f-8676-cefcc31b6e6c","Type":"ContainerDied","Data":"c22de03ac8c963cc300d233de1c8334cfdc65d1699252c9acfbf415608aba387"} Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.958123 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vcfbk" event={"ID":"f6fb66df-a4db-417b-90fc-c779777883be","Type":"ContainerStarted","Data":"724e4fa61153c196e5eca4a84647c2027c9b23b936751769834ef97f69b9e37b"} Oct 03 16:29:19 crc kubenswrapper[4744]: I1003 16:29:19.973294 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7v9wq"] Oct 03 16:29:19 crc kubenswrapper[4744]: W1003 16:29:19.980462 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d9cbc3a_4ccb_4e1e_8d56_3b457f5eebe2.slice/crio-c81d49e33d7c68110e0afe064d0a5103355c3a862514b6a596b42f3cbca936bd WatchSource:0}: Error finding container c81d49e33d7c68110e0afe064d0a5103355c3a862514b6a596b42f3cbca936bd: Status 404 returned error can't find the container with id c81d49e33d7c68110e0afe064d0a5103355c3a862514b6a596b42f3cbca936bd Oct 03 16:29:20 crc kubenswrapper[4744]: I1003 16:29:20.966633 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wmqbq" event={"ID":"86654f3a-97fa-456a-b651-93f8ec174725","Type":"ContainerStarted","Data":"feb4ece8af3b53cef3630af19817edc3bb30208e0144107f7b011a9d6a296df7"} Oct 03 16:29:20 crc kubenswrapper[4744]: I1003 16:29:20.970117 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l8gwc" event={"ID":"83f32f6a-6664-4e8f-8676-cefcc31b6e6c","Type":"ContainerStarted","Data":"504e9651fda516fab5ebac9ba581af307893e6679a20191f0754cfc05f32a917"} Oct 03 16:29:20 crc kubenswrapper[4744]: I1003 16:29:20.971877 4744 generic.go:334] "Generic (PLEG): container finished" podID="f6fb66df-a4db-417b-90fc-c779777883be" containerID="7c75e8c0a8e9256307cfc544aceb91f6b883a8e9aa156dd1c83b95489d6cd2fa" exitCode=0 Oct 03 16:29:20 crc kubenswrapper[4744]: I1003 16:29:20.971969 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vcfbk" event={"ID":"f6fb66df-a4db-417b-90fc-c779777883be","Type":"ContainerDied","Data":"7c75e8c0a8e9256307cfc544aceb91f6b883a8e9aa156dd1c83b95489d6cd2fa"} Oct 03 16:29:20 crc kubenswrapper[4744]: I1003 16:29:20.974039 4744 generic.go:334] "Generic (PLEG): container finished" podID="6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2" containerID="463cb752c807dd92d28f128a6aa8d2ae0dcb58859aac8334b7a11999a02c6fdd" exitCode=0 Oct 03 16:29:20 crc kubenswrapper[4744]: I1003 16:29:20.974089 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7v9wq" event={"ID":"6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2","Type":"ContainerDied","Data":"463cb752c807dd92d28f128a6aa8d2ae0dcb58859aac8334b7a11999a02c6fdd"} Oct 03 16:29:20 crc kubenswrapper[4744]: I1003 16:29:20.974111 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7v9wq" event={"ID":"6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2","Type":"ContainerStarted","Data":"c81d49e33d7c68110e0afe064d0a5103355c3a862514b6a596b42f3cbca936bd"} Oct 03 16:29:20 crc kubenswrapper[4744]: I1003 16:29:20.994355 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wmqbq" podStartSLOduration=2.380625811 podStartE2EDuration="4.994337017s" podCreationTimestamp="2025-10-03 16:29:16 +0000 UTC" firstStartedPulling="2025-10-03 16:29:17.922653512 +0000 UTC m=+284.202529408" lastFinishedPulling="2025-10-03 16:29:20.536364718 +0000 UTC m=+286.816240614" observedRunningTime="2025-10-03 16:29:20.99193956 +0000 UTC m=+287.271815476" watchObservedRunningTime="2025-10-03 16:29:20.994337017 +0000 UTC m=+287.274212913" Oct 03 16:29:21 crc kubenswrapper[4744]: I1003 16:29:21.035651 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l8gwc" podStartSLOduration=2.359057262 podStartE2EDuration="5.035628553s" podCreationTimestamp="2025-10-03 16:29:16 +0000 UTC" firstStartedPulling="2025-10-03 16:29:17.926060428 +0000 UTC m=+284.205936324" lastFinishedPulling="2025-10-03 16:29:20.602631719 +0000 UTC m=+286.882507615" observedRunningTime="2025-10-03 16:29:21.034678556 +0000 UTC m=+287.314554452" watchObservedRunningTime="2025-10-03 16:29:21.035628553 +0000 UTC m=+287.315504449" Oct 03 16:29:23 crc kubenswrapper[4744]: I1003 16:29:23.995201 4744 generic.go:334] "Generic (PLEG): container finished" podID="6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2" containerID="ef1877f48b57b7bc1b340e6dbffc480f7761829a15d5975ec1360904ed3a0535" exitCode=0 Oct 03 16:29:23 crc kubenswrapper[4744]: I1003 16:29:23.995266 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7v9wq" event={"ID":"6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2","Type":"ContainerDied","Data":"ef1877f48b57b7bc1b340e6dbffc480f7761829a15d5975ec1360904ed3a0535"} Oct 03 16:29:23 crc kubenswrapper[4744]: I1003 16:29:23.998300 4744 generic.go:334] "Generic (PLEG): container finished" podID="f6fb66df-a4db-417b-90fc-c779777883be" containerID="63cc8b4bb5d5e4773eabba7657e2ab70c02709da32393a917a1d10de06afda56" exitCode=0 Oct 03 16:29:23 crc kubenswrapper[4744]: I1003 16:29:23.998424 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vcfbk" event={"ID":"f6fb66df-a4db-417b-90fc-c779777883be","Type":"ContainerDied","Data":"63cc8b4bb5d5e4773eabba7657e2ab70c02709da32393a917a1d10de06afda56"} Oct 03 16:29:26 crc kubenswrapper[4744]: I1003 16:29:26.011557 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7v9wq" event={"ID":"6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2","Type":"ContainerStarted","Data":"c1050f9907ace761bfd179610594113b127516a612e2d878531f8bbdb14ac0a5"} Oct 03 16:29:26 crc kubenswrapper[4744]: I1003 16:29:26.014072 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vcfbk" event={"ID":"f6fb66df-a4db-417b-90fc-c779777883be","Type":"ContainerStarted","Data":"573fbf0c317f0e25120970c9fd9f9b1f0b59d696ef685b342a41bb6dad2b1400"} Oct 03 16:29:26 crc kubenswrapper[4744]: I1003 16:29:26.031218 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7v9wq" podStartSLOduration=3.135768089 podStartE2EDuration="7.03119538s" podCreationTimestamp="2025-10-03 16:29:19 +0000 UTC" firstStartedPulling="2025-10-03 16:29:20.976092172 +0000 UTC m=+287.255968058" lastFinishedPulling="2025-10-03 16:29:24.871519453 +0000 UTC m=+291.151395349" observedRunningTime="2025-10-03 16:29:26.028136164 +0000 UTC m=+292.308012090" watchObservedRunningTime="2025-10-03 16:29:26.03119538 +0000 UTC m=+292.311071266" Oct 03 16:29:26 crc kubenswrapper[4744]: I1003 16:29:26.048662 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vcfbk" podStartSLOduration=2.908774041 podStartE2EDuration="7.048635552s" podCreationTimestamp="2025-10-03 16:29:19 +0000 UTC" firstStartedPulling="2025-10-03 16:29:20.973571411 +0000 UTC m=+287.253447307" lastFinishedPulling="2025-10-03 16:29:25.113432922 +0000 UTC m=+291.393308818" observedRunningTime="2025-10-03 16:29:26.045068842 +0000 UTC m=+292.324944748" watchObservedRunningTime="2025-10-03 16:29:26.048635552 +0000 UTC m=+292.328511448" Oct 03 16:29:26 crc kubenswrapper[4744]: I1003 16:29:26.981668 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wmqbq" Oct 03 16:29:26 crc kubenswrapper[4744]: I1003 16:29:26.981753 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wmqbq" Oct 03 16:29:27 crc kubenswrapper[4744]: I1003 16:29:27.024411 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wmqbq" Oct 03 16:29:27 crc kubenswrapper[4744]: I1003 16:29:27.068587 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wmqbq" Oct 03 16:29:27 crc kubenswrapper[4744]: I1003 16:29:27.199531 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l8gwc" Oct 03 16:29:27 crc kubenswrapper[4744]: I1003 16:29:27.199590 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l8gwc" Oct 03 16:29:27 crc kubenswrapper[4744]: I1003 16:29:27.240706 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l8gwc" Oct 03 16:29:28 crc kubenswrapper[4744]: I1003 16:29:28.074711 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l8gwc" Oct 03 16:29:29 crc kubenswrapper[4744]: I1003 16:29:29.380647 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vcfbk" Oct 03 16:29:29 crc kubenswrapper[4744]: I1003 16:29:29.380711 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vcfbk" Oct 03 16:29:29 crc kubenswrapper[4744]: I1003 16:29:29.436767 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vcfbk" Oct 03 16:29:29 crc kubenswrapper[4744]: I1003 16:29:29.559658 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7v9wq" Oct 03 16:29:29 crc kubenswrapper[4744]: I1003 16:29:29.559742 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7v9wq" Oct 03 16:29:29 crc kubenswrapper[4744]: I1003 16:29:29.602094 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7v9wq" Oct 03 16:29:30 crc kubenswrapper[4744]: I1003 16:29:30.085029 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vcfbk" Oct 03 16:29:30 crc kubenswrapper[4744]: I1003 16:29:30.089439 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7v9wq" Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.155649 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt"] Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.158060 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.164029 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.164561 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.169454 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt"] Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.197965 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-config-volume\") pod \"collect-profiles-29325150-4ktxt\" (UID: \"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.198132 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlnb9\" (UniqueName: \"kubernetes.io/projected/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-kube-api-access-nlnb9\") pod \"collect-profiles-29325150-4ktxt\" (UID: \"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.198211 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-secret-volume\") pod \"collect-profiles-29325150-4ktxt\" (UID: \"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.298878 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-config-volume\") pod \"collect-profiles-29325150-4ktxt\" (UID: \"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.298946 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlnb9\" (UniqueName: \"kubernetes.io/projected/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-kube-api-access-nlnb9\") pod \"collect-profiles-29325150-4ktxt\" (UID: \"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.298984 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-secret-volume\") pod \"collect-profiles-29325150-4ktxt\" (UID: \"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.299857 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-config-volume\") pod \"collect-profiles-29325150-4ktxt\" (UID: \"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.307715 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-secret-volume\") pod \"collect-profiles-29325150-4ktxt\" (UID: \"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.315550 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlnb9\" (UniqueName: \"kubernetes.io/projected/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-kube-api-access-nlnb9\") pod \"collect-profiles-29325150-4ktxt\" (UID: \"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.484955 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" Oct 03 16:30:00 crc kubenswrapper[4744]: I1003 16:30:00.736983 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt"] Oct 03 16:30:01 crc kubenswrapper[4744]: I1003 16:30:01.232302 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" event={"ID":"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f","Type":"ContainerStarted","Data":"460cfb11a0f81225424d3d509a800349764ca50030e40fbccac077bce1c5a4fc"} Oct 03 16:30:01 crc kubenswrapper[4744]: I1003 16:30:01.232733 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" event={"ID":"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f","Type":"ContainerStarted","Data":"eeba7c7674e1b61b1d95e118c5ea3e9dc83d2763fd2a0da4441dc64d58917b69"} Oct 03 16:30:01 crc kubenswrapper[4744]: I1003 16:30:01.251910 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" podStartSLOduration=1.251881244 podStartE2EDuration="1.251881244s" podCreationTimestamp="2025-10-03 16:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:30:01.251796481 +0000 UTC m=+327.531672447" watchObservedRunningTime="2025-10-03 16:30:01.251881244 +0000 UTC m=+327.531757140" Oct 03 16:30:02 crc kubenswrapper[4744]: I1003 16:30:02.243574 4744 generic.go:334] "Generic (PLEG): container finished" podID="bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f" containerID="460cfb11a0f81225424d3d509a800349764ca50030e40fbccac077bce1c5a4fc" exitCode=0 Oct 03 16:30:02 crc kubenswrapper[4744]: I1003 16:30:02.243667 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" event={"ID":"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f","Type":"ContainerDied","Data":"460cfb11a0f81225424d3d509a800349764ca50030e40fbccac077bce1c5a4fc"} Oct 03 16:30:03 crc kubenswrapper[4744]: I1003 16:30:03.477244 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" Oct 03 16:30:03 crc kubenswrapper[4744]: I1003 16:30:03.647799 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-config-volume\") pod \"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f\" (UID: \"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f\") " Oct 03 16:30:03 crc kubenswrapper[4744]: I1003 16:30:03.647895 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlnb9\" (UniqueName: \"kubernetes.io/projected/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-kube-api-access-nlnb9\") pod \"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f\" (UID: \"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f\") " Oct 03 16:30:03 crc kubenswrapper[4744]: I1003 16:30:03.648106 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-secret-volume\") pod \"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f\" (UID: \"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f\") " Oct 03 16:30:03 crc kubenswrapper[4744]: I1003 16:30:03.648770 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-config-volume" (OuterVolumeSpecName: "config-volume") pod "bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f" (UID: "bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:30:03 crc kubenswrapper[4744]: I1003 16:30:03.655553 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-kube-api-access-nlnb9" (OuterVolumeSpecName: "kube-api-access-nlnb9") pod "bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f" (UID: "bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f"). InnerVolumeSpecName "kube-api-access-nlnb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:30:03 crc kubenswrapper[4744]: I1003 16:30:03.656443 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f" (UID: "bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:30:03 crc kubenswrapper[4744]: I1003 16:30:03.749414 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 16:30:03 crc kubenswrapper[4744]: I1003 16:30:03.749456 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 16:30:03 crc kubenswrapper[4744]: I1003 16:30:03.749468 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlnb9\" (UniqueName: \"kubernetes.io/projected/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f-kube-api-access-nlnb9\") on node \"crc\" DevicePath \"\"" Oct 03 16:30:04 crc kubenswrapper[4744]: I1003 16:30:04.256664 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" event={"ID":"bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f","Type":"ContainerDied","Data":"eeba7c7674e1b61b1d95e118c5ea3e9dc83d2763fd2a0da4441dc64d58917b69"} Oct 03 16:30:04 crc kubenswrapper[4744]: I1003 16:30:04.256719 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eeba7c7674e1b61b1d95e118c5ea3e9dc83d2763fd2a0da4441dc64d58917b69" Oct 03 16:30:04 crc kubenswrapper[4744]: I1003 16:30:04.256753 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt" Oct 03 16:30:34 crc kubenswrapper[4744]: I1003 16:30:34.668285 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:30:34 crc kubenswrapper[4744]: I1003 16:30:34.669103 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:31:04 crc kubenswrapper[4744]: I1003 16:31:04.668386 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:31:04 crc kubenswrapper[4744]: I1003 16:31:04.669274 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:31:25 crc kubenswrapper[4744]: I1003 16:31:25.983484 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-jwwfx"] Oct 03 16:31:25 crc kubenswrapper[4744]: E1003 16:31:25.984628 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f" containerName="collect-profiles" Oct 03 16:31:25 crc kubenswrapper[4744]: I1003 16:31:25.985201 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f" containerName="collect-profiles" Oct 03 16:31:25 crc kubenswrapper[4744]: I1003 16:31:25.985323 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f" containerName="collect-profiles" Oct 03 16:31:25 crc kubenswrapper[4744]: I1003 16:31:25.985803 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:25 crc kubenswrapper[4744]: I1003 16:31:25.998067 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-jwwfx"] Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.091756 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2cefe71f-8d8c-440b-8763-9bae83a10b4b-bound-sa-token\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.091814 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2cefe71f-8d8c-440b-8763-9bae83a10b4b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.091844 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.091868 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2cefe71f-8d8c-440b-8763-9bae83a10b4b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.091928 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2cefe71f-8d8c-440b-8763-9bae83a10b4b-trusted-ca\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.091960 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2cefe71f-8d8c-440b-8763-9bae83a10b4b-registry-certificates\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.092019 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52n2s\" (UniqueName: \"kubernetes.io/projected/2cefe71f-8d8c-440b-8763-9bae83a10b4b-kube-api-access-52n2s\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.092232 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2cefe71f-8d8c-440b-8763-9bae83a10b4b-registry-tls\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.135227 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.193871 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2cefe71f-8d8c-440b-8763-9bae83a10b4b-trusted-ca\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.193948 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2cefe71f-8d8c-440b-8763-9bae83a10b4b-registry-certificates\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.193972 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52n2s\" (UniqueName: \"kubernetes.io/projected/2cefe71f-8d8c-440b-8763-9bae83a10b4b-kube-api-access-52n2s\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.194000 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2cefe71f-8d8c-440b-8763-9bae83a10b4b-registry-tls\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.194036 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2cefe71f-8d8c-440b-8763-9bae83a10b4b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.194061 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2cefe71f-8d8c-440b-8763-9bae83a10b4b-bound-sa-token\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.194083 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2cefe71f-8d8c-440b-8763-9bae83a10b4b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.195014 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2cefe71f-8d8c-440b-8763-9bae83a10b4b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.196094 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2cefe71f-8d8c-440b-8763-9bae83a10b4b-trusted-ca\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.196094 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2cefe71f-8d8c-440b-8763-9bae83a10b4b-registry-certificates\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.201790 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2cefe71f-8d8c-440b-8763-9bae83a10b4b-registry-tls\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.202881 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2cefe71f-8d8c-440b-8763-9bae83a10b4b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.210360 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2cefe71f-8d8c-440b-8763-9bae83a10b4b-bound-sa-token\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.211314 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52n2s\" (UniqueName: \"kubernetes.io/projected/2cefe71f-8d8c-440b-8763-9bae83a10b4b-kube-api-access-52n2s\") pod \"image-registry-66df7c8f76-jwwfx\" (UID: \"2cefe71f-8d8c-440b-8763-9bae83a10b4b\") " pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.310114 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.518319 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-jwwfx"] Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.811752 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" event={"ID":"2cefe71f-8d8c-440b-8763-9bae83a10b4b","Type":"ContainerStarted","Data":"ae763eaf38bea929f507888e8b5d4b15454d69dd96e8d35dcc65904beae57e47"} Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.811896 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.811921 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" event={"ID":"2cefe71f-8d8c-440b-8763-9bae83a10b4b","Type":"ContainerStarted","Data":"7c733255cecb292f62ff72c772ecfca62221a6037f1933c0153ff0caff8e7820"} Oct 03 16:31:26 crc kubenswrapper[4744]: I1003 16:31:26.839044 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" podStartSLOduration=1.8390170829999999 podStartE2EDuration="1.839017083s" podCreationTimestamp="2025-10-03 16:31:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:31:26.836566396 +0000 UTC m=+413.116442292" watchObservedRunningTime="2025-10-03 16:31:26.839017083 +0000 UTC m=+413.118892979" Oct 03 16:31:34 crc kubenswrapper[4744]: I1003 16:31:34.669482 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:31:34 crc kubenswrapper[4744]: I1003 16:31:34.670643 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:31:34 crc kubenswrapper[4744]: I1003 16:31:34.670715 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:31:34 crc kubenswrapper[4744]: I1003 16:31:34.671520 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ebd68629db9100ac91296997be91b152693b5453b3c435c930ae88e90650ca93"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 16:31:34 crc kubenswrapper[4744]: I1003 16:31:34.671785 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://ebd68629db9100ac91296997be91b152693b5453b3c435c930ae88e90650ca93" gracePeriod=600 Oct 03 16:31:34 crc kubenswrapper[4744]: I1003 16:31:34.872406 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="ebd68629db9100ac91296997be91b152693b5453b3c435c930ae88e90650ca93" exitCode=0 Oct 03 16:31:34 crc kubenswrapper[4744]: I1003 16:31:34.872558 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"ebd68629db9100ac91296997be91b152693b5453b3c435c930ae88e90650ca93"} Oct 03 16:31:34 crc kubenswrapper[4744]: I1003 16:31:34.873109 4744 scope.go:117] "RemoveContainer" containerID="f16754cc2f88747cb419b8e33e63b8909d7096602826cd3bd45a1eb4da63939e" Oct 03 16:31:35 crc kubenswrapper[4744]: I1003 16:31:35.882680 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"c5619e90872156a7cc8093cc5e7ceeb12da127b0d8d653ccf7ad40c5326f1a97"} Oct 03 16:31:46 crc kubenswrapper[4744]: I1003 16:31:46.316176 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-jwwfx" Oct 03 16:31:46 crc kubenswrapper[4744]: I1003 16:31:46.382337 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-6tzpn"] Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.430623 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" podUID="cc342582-3925-4af1-b9b8-8e1333af091c" containerName="registry" containerID="cri-o://b9e2148bb457a4bd1e40b0a8cd550f2d60b5d1bb008524d78e75de3b1cd3550d" gracePeriod=30 Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.830077 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.864065 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cc342582-3925-4af1-b9b8-8e1333af091c-registry-certificates\") pod \"cc342582-3925-4af1-b9b8-8e1333af091c\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.864274 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"cc342582-3925-4af1-b9b8-8e1333af091c\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.864327 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cc342582-3925-4af1-b9b8-8e1333af091c-ca-trust-extracted\") pod \"cc342582-3925-4af1-b9b8-8e1333af091c\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.864388 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfzdj\" (UniqueName: \"kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-kube-api-access-sfzdj\") pod \"cc342582-3925-4af1-b9b8-8e1333af091c\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.864420 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-registry-tls\") pod \"cc342582-3925-4af1-b9b8-8e1333af091c\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.864447 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc342582-3925-4af1-b9b8-8e1333af091c-trusted-ca\") pod \"cc342582-3925-4af1-b9b8-8e1333af091c\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.864471 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cc342582-3925-4af1-b9b8-8e1333af091c-installation-pull-secrets\") pod \"cc342582-3925-4af1-b9b8-8e1333af091c\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.864516 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-bound-sa-token\") pod \"cc342582-3925-4af1-b9b8-8e1333af091c\" (UID: \"cc342582-3925-4af1-b9b8-8e1333af091c\") " Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.865448 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc342582-3925-4af1-b9b8-8e1333af091c-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "cc342582-3925-4af1-b9b8-8e1333af091c" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.865481 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc342582-3925-4af1-b9b8-8e1333af091c-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "cc342582-3925-4af1-b9b8-8e1333af091c" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.872426 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc342582-3925-4af1-b9b8-8e1333af091c-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "cc342582-3925-4af1-b9b8-8e1333af091c" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.873221 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-kube-api-access-sfzdj" (OuterVolumeSpecName: "kube-api-access-sfzdj") pod "cc342582-3925-4af1-b9b8-8e1333af091c" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c"). InnerVolumeSpecName "kube-api-access-sfzdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.873521 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "cc342582-3925-4af1-b9b8-8e1333af091c" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.873822 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "cc342582-3925-4af1-b9b8-8e1333af091c" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.878478 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "cc342582-3925-4af1-b9b8-8e1333af091c" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.896591 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc342582-3925-4af1-b9b8-8e1333af091c-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "cc342582-3925-4af1-b9b8-8e1333af091c" (UID: "cc342582-3925-4af1-b9b8-8e1333af091c"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.966753 4744 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cc342582-3925-4af1-b9b8-8e1333af091c-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.966809 4744 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cc342582-3925-4af1-b9b8-8e1333af091c-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.966823 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfzdj\" (UniqueName: \"kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-kube-api-access-sfzdj\") on node \"crc\" DevicePath \"\"" Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.966836 4744 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.966851 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc342582-3925-4af1-b9b8-8e1333af091c-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.966866 4744 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cc342582-3925-4af1-b9b8-8e1333af091c-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 03 16:32:11 crc kubenswrapper[4744]: I1003 16:32:11.966878 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cc342582-3925-4af1-b9b8-8e1333af091c-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 03 16:32:12 crc kubenswrapper[4744]: I1003 16:32:12.094128 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc342582-3925-4af1-b9b8-8e1333af091c" containerID="b9e2148bb457a4bd1e40b0a8cd550f2d60b5d1bb008524d78e75de3b1cd3550d" exitCode=0 Oct 03 16:32:12 crc kubenswrapper[4744]: I1003 16:32:12.094178 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" event={"ID":"cc342582-3925-4af1-b9b8-8e1333af091c","Type":"ContainerDied","Data":"b9e2148bb457a4bd1e40b0a8cd550f2d60b5d1bb008524d78e75de3b1cd3550d"} Oct 03 16:32:12 crc kubenswrapper[4744]: I1003 16:32:12.094214 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" event={"ID":"cc342582-3925-4af1-b9b8-8e1333af091c","Type":"ContainerDied","Data":"f4bb15b9977ff10a5b109cde5d357b0684eec284957bd238e917118e253694b5"} Oct 03 16:32:12 crc kubenswrapper[4744]: I1003 16:32:12.094235 4744 scope.go:117] "RemoveContainer" containerID="b9e2148bb457a4bd1e40b0a8cd550f2d60b5d1bb008524d78e75de3b1cd3550d" Oct 03 16:32:12 crc kubenswrapper[4744]: I1003 16:32:12.094350 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-6tzpn" Oct 03 16:32:12 crc kubenswrapper[4744]: I1003 16:32:12.119445 4744 scope.go:117] "RemoveContainer" containerID="b9e2148bb457a4bd1e40b0a8cd550f2d60b5d1bb008524d78e75de3b1cd3550d" Oct 03 16:32:12 crc kubenswrapper[4744]: E1003 16:32:12.122190 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9e2148bb457a4bd1e40b0a8cd550f2d60b5d1bb008524d78e75de3b1cd3550d\": container with ID starting with b9e2148bb457a4bd1e40b0a8cd550f2d60b5d1bb008524d78e75de3b1cd3550d not found: ID does not exist" containerID="b9e2148bb457a4bd1e40b0a8cd550f2d60b5d1bb008524d78e75de3b1cd3550d" Oct 03 16:32:12 crc kubenswrapper[4744]: I1003 16:32:12.122249 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e2148bb457a4bd1e40b0a8cd550f2d60b5d1bb008524d78e75de3b1cd3550d"} err="failed to get container status \"b9e2148bb457a4bd1e40b0a8cd550f2d60b5d1bb008524d78e75de3b1cd3550d\": rpc error: code = NotFound desc = could not find container \"b9e2148bb457a4bd1e40b0a8cd550f2d60b5d1bb008524d78e75de3b1cd3550d\": container with ID starting with b9e2148bb457a4bd1e40b0a8cd550f2d60b5d1bb008524d78e75de3b1cd3550d not found: ID does not exist" Oct 03 16:32:12 crc kubenswrapper[4744]: I1003 16:32:12.128346 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-6tzpn"] Oct 03 16:32:12 crc kubenswrapper[4744]: I1003 16:32:12.131615 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-6tzpn"] Oct 03 16:32:12 crc kubenswrapper[4744]: I1003 16:32:12.901552 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc342582-3925-4af1-b9b8-8e1333af091c" path="/var/lib/kubelet/pods/cc342582-3925-4af1-b9b8-8e1333af091c/volumes" Oct 03 16:34:04 crc kubenswrapper[4744]: I1003 16:34:04.669173 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:34:04 crc kubenswrapper[4744]: I1003 16:34:04.670119 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:34:34 crc kubenswrapper[4744]: I1003 16:34:34.668796 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:34:34 crc kubenswrapper[4744]: I1003 16:34:34.669898 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.217680 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-xp59p"] Oct 03 16:34:50 crc kubenswrapper[4744]: E1003 16:34:50.218712 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc342582-3925-4af1-b9b8-8e1333af091c" containerName="registry" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.218732 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc342582-3925-4af1-b9b8-8e1333af091c" containerName="registry" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.218875 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc342582-3925-4af1-b9b8-8e1333af091c" containerName="registry" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.219460 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-xp59p" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.224894 4744 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-f26xj" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.225084 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-m8bhx"] Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.225313 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.226258 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-m8bhx" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.228154 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.230266 4744 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-4n9cl" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.236823 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-xp59p"] Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.250185 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xjmdh"] Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.251239 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xjmdh" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.257590 4744 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-97qdq" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.268609 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-m8bhx"] Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.284380 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xjmdh"] Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.302441 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s87c\" (UniqueName: \"kubernetes.io/projected/862c46f4-2e50-4556-8405-bfcdb337ed95-kube-api-access-8s87c\") pod \"cert-manager-cainjector-7f985d654d-xp59p\" (UID: \"862c46f4-2e50-4556-8405-bfcdb337ed95\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-xp59p" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.403372 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s87c\" (UniqueName: \"kubernetes.io/projected/862c46f4-2e50-4556-8405-bfcdb337ed95-kube-api-access-8s87c\") pod \"cert-manager-cainjector-7f985d654d-xp59p\" (UID: \"862c46f4-2e50-4556-8405-bfcdb337ed95\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-xp59p" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.403450 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwchm\" (UniqueName: \"kubernetes.io/projected/15fdf250-5959-4502-b7be-4d7af4eff05d-kube-api-access-gwchm\") pod \"cert-manager-5b446d88c5-m8bhx\" (UID: \"15fdf250-5959-4502-b7be-4d7af4eff05d\") " pod="cert-manager/cert-manager-5b446d88c5-m8bhx" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.403568 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dt7b\" (UniqueName: \"kubernetes.io/projected/f62bcdb5-0b26-4be1-af80-c413f76ce50c-kube-api-access-5dt7b\") pod \"cert-manager-webhook-5655c58dd6-xjmdh\" (UID: \"f62bcdb5-0b26-4be1-af80-c413f76ce50c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xjmdh" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.423277 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s87c\" (UniqueName: \"kubernetes.io/projected/862c46f4-2e50-4556-8405-bfcdb337ed95-kube-api-access-8s87c\") pod \"cert-manager-cainjector-7f985d654d-xp59p\" (UID: \"862c46f4-2e50-4556-8405-bfcdb337ed95\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-xp59p" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.504872 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwchm\" (UniqueName: \"kubernetes.io/projected/15fdf250-5959-4502-b7be-4d7af4eff05d-kube-api-access-gwchm\") pod \"cert-manager-5b446d88c5-m8bhx\" (UID: \"15fdf250-5959-4502-b7be-4d7af4eff05d\") " pod="cert-manager/cert-manager-5b446d88c5-m8bhx" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.504984 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dt7b\" (UniqueName: \"kubernetes.io/projected/f62bcdb5-0b26-4be1-af80-c413f76ce50c-kube-api-access-5dt7b\") pod \"cert-manager-webhook-5655c58dd6-xjmdh\" (UID: \"f62bcdb5-0b26-4be1-af80-c413f76ce50c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xjmdh" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.529519 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dt7b\" (UniqueName: \"kubernetes.io/projected/f62bcdb5-0b26-4be1-af80-c413f76ce50c-kube-api-access-5dt7b\") pod \"cert-manager-webhook-5655c58dd6-xjmdh\" (UID: \"f62bcdb5-0b26-4be1-af80-c413f76ce50c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xjmdh" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.537278 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwchm\" (UniqueName: \"kubernetes.io/projected/15fdf250-5959-4502-b7be-4d7af4eff05d-kube-api-access-gwchm\") pod \"cert-manager-5b446d88c5-m8bhx\" (UID: \"15fdf250-5959-4502-b7be-4d7af4eff05d\") " pod="cert-manager/cert-manager-5b446d88c5-m8bhx" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.549070 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-xp59p" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.560636 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-m8bhx" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.577419 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xjmdh" Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.807878 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-xp59p"] Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.858466 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.882483 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-m8bhx"] Oct 03 16:34:50 crc kubenswrapper[4744]: I1003 16:34:50.901313 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xjmdh"] Oct 03 16:34:50 crc kubenswrapper[4744]: W1003 16:34:50.906568 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf62bcdb5_0b26_4be1_af80_c413f76ce50c.slice/crio-014a11aafa14540897721b47c98c5fa503b6f77db1937d8829de0c47c0c7fa44 WatchSource:0}: Error finding container 014a11aafa14540897721b47c98c5fa503b6f77db1937d8829de0c47c0c7fa44: Status 404 returned error can't find the container with id 014a11aafa14540897721b47c98c5fa503b6f77db1937d8829de0c47c0c7fa44 Oct 03 16:34:51 crc kubenswrapper[4744]: I1003 16:34:51.170607 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-m8bhx" event={"ID":"15fdf250-5959-4502-b7be-4d7af4eff05d","Type":"ContainerStarted","Data":"9576719af3e143ba2a0bcb0b13ce499bbed23a1419de244e6c59a3cdc4ab2ebc"} Oct 03 16:34:51 crc kubenswrapper[4744]: I1003 16:34:51.171693 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-xp59p" event={"ID":"862c46f4-2e50-4556-8405-bfcdb337ed95","Type":"ContainerStarted","Data":"4d69909d3fc58a224e4f00c66bfbf523b44381e5b31883f3d6b972985989ee07"} Oct 03 16:34:51 crc kubenswrapper[4744]: I1003 16:34:51.172989 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xjmdh" event={"ID":"f62bcdb5-0b26-4be1-af80-c413f76ce50c","Type":"ContainerStarted","Data":"014a11aafa14540897721b47c98c5fa503b6f77db1937d8829de0c47c0c7fa44"} Oct 03 16:34:57 crc kubenswrapper[4744]: I1003 16:34:57.209120 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-m8bhx" event={"ID":"15fdf250-5959-4502-b7be-4d7af4eff05d","Type":"ContainerStarted","Data":"9df965950d424599d391a815f374049fba0cb837abca5338ea5e8a81e0e346fc"} Oct 03 16:34:57 crc kubenswrapper[4744]: I1003 16:34:57.211451 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-xp59p" event={"ID":"862c46f4-2e50-4556-8405-bfcdb337ed95","Type":"ContainerStarted","Data":"2d3562e27c7cb5a56ecf747673ced6c8317f087d069b6fe0f3fe5e0372161501"} Oct 03 16:34:57 crc kubenswrapper[4744]: I1003 16:34:57.212976 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xjmdh" event={"ID":"f62bcdb5-0b26-4be1-af80-c413f76ce50c","Type":"ContainerStarted","Data":"500167da646be0a78b9595fb8295fabd2b00119947ef4b6aade7184ddd06e79f"} Oct 03 16:34:57 crc kubenswrapper[4744]: I1003 16:34:57.213121 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-xjmdh" Oct 03 16:34:57 crc kubenswrapper[4744]: I1003 16:34:57.235014 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-m8bhx" podStartSLOduration=1.435135317 podStartE2EDuration="7.234992936s" podCreationTimestamp="2025-10-03 16:34:50 +0000 UTC" firstStartedPulling="2025-10-03 16:34:50.889206054 +0000 UTC m=+617.169081950" lastFinishedPulling="2025-10-03 16:34:56.689063673 +0000 UTC m=+622.968939569" observedRunningTime="2025-10-03 16:34:57.229041667 +0000 UTC m=+623.508917573" watchObservedRunningTime="2025-10-03 16:34:57.234992936 +0000 UTC m=+623.514868832" Oct 03 16:34:57 crc kubenswrapper[4744]: I1003 16:34:57.249553 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-xjmdh" podStartSLOduration=1.470533425 podStartE2EDuration="7.249486134s" podCreationTimestamp="2025-10-03 16:34:50 +0000 UTC" firstStartedPulling="2025-10-03 16:34:50.90996661 +0000 UTC m=+617.189842506" lastFinishedPulling="2025-10-03 16:34:56.688919319 +0000 UTC m=+622.968795215" observedRunningTime="2025-10-03 16:34:57.248484517 +0000 UTC m=+623.528360433" watchObservedRunningTime="2025-10-03 16:34:57.249486134 +0000 UTC m=+623.529362020" Oct 03 16:34:57 crc kubenswrapper[4744]: I1003 16:34:57.275098 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-xp59p" podStartSLOduration=1.378120363 podStartE2EDuration="7.275080889s" podCreationTimestamp="2025-10-03 16:34:50 +0000 UTC" firstStartedPulling="2025-10-03 16:34:50.858120733 +0000 UTC m=+617.137996629" lastFinishedPulling="2025-10-03 16:34:56.755081259 +0000 UTC m=+623.034957155" observedRunningTime="2025-10-03 16:34:57.271464522 +0000 UTC m=+623.551340418" watchObservedRunningTime="2025-10-03 16:34:57.275080889 +0000 UTC m=+623.554956785" Oct 03 16:35:00 crc kubenswrapper[4744]: I1003 16:35:00.851408 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jw5rc"] Oct 03 16:35:00 crc kubenswrapper[4744]: I1003 16:35:00.852326 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovn-controller" containerID="cri-o://aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0" gracePeriod=30 Oct 03 16:35:00 crc kubenswrapper[4744]: I1003 16:35:00.852439 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="nbdb" containerID="cri-o://4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64" gracePeriod=30 Oct 03 16:35:00 crc kubenswrapper[4744]: I1003 16:35:00.852561 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c" gracePeriod=30 Oct 03 16:35:00 crc kubenswrapper[4744]: I1003 16:35:00.852485 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="northd" containerID="cri-o://aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d" gracePeriod=30 Oct 03 16:35:00 crc kubenswrapper[4744]: I1003 16:35:00.852684 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovn-acl-logging" containerID="cri-o://88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558" gracePeriod=30 Oct 03 16:35:00 crc kubenswrapper[4744]: I1003 16:35:00.852744 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="kube-rbac-proxy-node" containerID="cri-o://a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603" gracePeriod=30 Oct 03 16:35:00 crc kubenswrapper[4744]: I1003 16:35:00.852789 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="sbdb" containerID="cri-o://1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe" gracePeriod=30 Oct 03 16:35:00 crc kubenswrapper[4744]: I1003 16:35:00.906318 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" containerID="cri-o://8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0" gracePeriod=30 Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.180316 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovnkube-controller/3.log" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.184880 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovn-acl-logging/0.log" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.185785 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovn-controller/0.log" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.186467 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.242852 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ppskd"] Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.243255 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovn-acl-logging" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243279 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovn-acl-logging" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.243296 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="nbdb" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243304 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="nbdb" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.243313 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243321 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.243329 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="sbdb" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243336 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="sbdb" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.243348 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="kube-rbac-proxy-node" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243356 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="kube-rbac-proxy-node" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.243364 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243371 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.243379 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243388 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.243400 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="kubecfg-setup" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243410 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="kubecfg-setup" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.243421 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243428 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.243438 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovn-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243444 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovn-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.243454 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="kube-rbac-proxy-ovn-metrics" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243461 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="kube-rbac-proxy-ovn-metrics" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.243473 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="northd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243480 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="northd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243623 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243635 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="sbdb" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243643 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243654 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243666 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovn-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243677 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovn-acl-logging" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243688 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="kube-rbac-proxy-ovn-metrics" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243702 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="kube-rbac-proxy-node" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243712 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="nbdb" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243722 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="northd" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.243830 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243840 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.243941 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.246973 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerName="ovnkube-controller" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.256302 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovnkube-controller/3.log" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.258627 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.262681 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovn-acl-logging/0.log" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.264879 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-jw5rc_3436425f-b9cf-4302-b5ad-ccb92d597aed/ovn-controller/0.log" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265612 4744 generic.go:334] "Generic (PLEG): container finished" podID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerID="8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0" exitCode=0 Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265648 4744 generic.go:334] "Generic (PLEG): container finished" podID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerID="1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe" exitCode=0 Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265655 4744 generic.go:334] "Generic (PLEG): container finished" podID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerID="4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64" exitCode=0 Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265662 4744 generic.go:334] "Generic (PLEG): container finished" podID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerID="aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d" exitCode=0 Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265670 4744 generic.go:334] "Generic (PLEG): container finished" podID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerID="c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c" exitCode=0 Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265677 4744 generic.go:334] "Generic (PLEG): container finished" podID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerID="a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603" exitCode=0 Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265685 4744 generic.go:334] "Generic (PLEG): container finished" podID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerID="88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558" exitCode=143 Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265692 4744 generic.go:334] "Generic (PLEG): container finished" podID="3436425f-b9cf-4302-b5ad-ccb92d597aed" containerID="aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0" exitCode=143 Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265729 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265776 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerDied","Data":"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265809 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerDied","Data":"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265823 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerDied","Data":"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265837 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerDied","Data":"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265851 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerDied","Data":"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265864 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerDied","Data":"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265880 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265895 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265904 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265911 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265917 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265923 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265929 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265936 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265942 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265949 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerDied","Data":"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265958 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265965 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265970 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265976 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265981 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265987 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265993 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.265998 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266003 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266008 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266015 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerDied","Data":"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266023 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266028 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266034 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266040 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266048 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266054 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266061 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266068 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266075 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266083 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266093 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jw5rc" event={"ID":"3436425f-b9cf-4302-b5ad-ccb92d597aed","Type":"ContainerDied","Data":"729b5896f0a8615cceb963cc8a7da51f07b98eaad1f1e4c70ded6c519d99f2f6"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266102 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266109 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266115 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266120 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266126 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266131 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266137 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266143 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266150 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266157 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.266178 4744 scope.go:117] "RemoveContainer" containerID="8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.270804 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vqz6q_84192d79-366a-453f-b70b-aefa4537ec4c/kube-multus/2.log" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.271639 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vqz6q_84192d79-366a-453f-b70b-aefa4537ec4c/kube-multus/1.log" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.271706 4744 generic.go:334] "Generic (PLEG): container finished" podID="84192d79-366a-453f-b70b-aefa4537ec4c" containerID="b9031bf515dd80acd90501c31a1f575672257c7875c2e6b608cf36a4e51744a4" exitCode=2 Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.271758 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vqz6q" event={"ID":"84192d79-366a-453f-b70b-aefa4537ec4c","Type":"ContainerDied","Data":"b9031bf515dd80acd90501c31a1f575672257c7875c2e6b608cf36a4e51744a4"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.271805 4744 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d3888e7ce3b4e4e7967e3c7ec2e5ef8b7bee43a7742087ca6ea6f44a3d72e45c"} Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.272512 4744 scope.go:117] "RemoveContainer" containerID="b9031bf515dd80acd90501c31a1f575672257c7875c2e6b608cf36a4e51744a4" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.272938 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-vqz6q_openshift-multus(84192d79-366a-453f-b70b-aefa4537ec4c)\"" pod="openshift-multus/multus-vqz6q" podUID="84192d79-366a-453f-b70b-aefa4537ec4c" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.293619 4744 scope.go:117] "RemoveContainer" containerID="e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.317588 4744 scope.go:117] "RemoveContainer" containerID="1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.332412 4744 scope.go:117] "RemoveContainer" containerID="4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.346012 4744 scope.go:117] "RemoveContainer" containerID="aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.361229 4744 scope.go:117] "RemoveContainer" containerID="c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375122 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-var-lib-openvswitch\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375168 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-run-netns\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375233 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovnkube-config\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375256 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovnkube-script-lib\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375275 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovn-node-metrics-cert\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375294 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-env-overrides\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375358 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxw7s\" (UniqueName: \"kubernetes.io/projected/3436425f-b9cf-4302-b5ad-ccb92d597aed-kube-api-access-nxw7s\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375381 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-node-log\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375399 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-slash\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375424 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-log-socket\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375440 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-etc-openvswitch\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375458 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-cni-bin\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375493 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-var-lib-cni-networks-ovn-kubernetes\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375543 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-run-ovn-kubernetes\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375568 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-ovn\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375589 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-systemd\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375613 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-cni-netd\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375645 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-systemd-units\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375665 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-kubelet\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375681 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-node-log" (OuterVolumeSpecName: "node-log") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375701 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-openvswitch\") pod \"3436425f-b9cf-4302-b5ad-ccb92d597aed\" (UID: \"3436425f-b9cf-4302-b5ad-ccb92d597aed\") " Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375908 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-cni-netd\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375954 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-run-ovn\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375989 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376012 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-ovnkube-config\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376047 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-run-netns\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376077 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-systemd-units\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376101 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk5s6\" (UniqueName: \"kubernetes.io/projected/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-kube-api-access-xk5s6\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376125 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-node-log\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376149 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-run-openvswitch\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376176 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-var-lib-openvswitch\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375704 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-slash" (OuterVolumeSpecName: "host-slash") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375681 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375718 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-log-socket" (OuterVolumeSpecName: "log-socket") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375732 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375732 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.375752 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376245 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-etc-openvswitch\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376112 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376153 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376200 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376220 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376241 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376263 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376689 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376820 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376824 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-ovnkube-script-lib\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376874 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-env-overrides\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376912 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-kubelet\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376935 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-slash\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.376979 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-cni-bin\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377009 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-ovn-node-metrics-cert\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377043 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-run-ovn-kubernetes\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377075 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-run-systemd\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377100 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-log-socket\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377188 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377213 4744 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377227 4744 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-node-log\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377214 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377239 4744 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-slash\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377284 4744 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-log-socket\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377304 4744 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377317 4744 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377332 4744 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377349 4744 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377368 4744 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377381 4744 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377393 4744 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377407 4744 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377420 4744 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377435 4744 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.377779 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.378098 4744 scope.go:117] "RemoveContainer" containerID="a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.383387 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3436425f-b9cf-4302-b5ad-ccb92d597aed-kube-api-access-nxw7s" (OuterVolumeSpecName: "kube-api-access-nxw7s") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "kube-api-access-nxw7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.383759 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.394078 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "3436425f-b9cf-4302-b5ad-ccb92d597aed" (UID: "3436425f-b9cf-4302-b5ad-ccb92d597aed"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.395333 4744 scope.go:117] "RemoveContainer" containerID="88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.411620 4744 scope.go:117] "RemoveContainer" containerID="aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.425893 4744 scope.go:117] "RemoveContainer" containerID="a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.442508 4744 scope.go:117] "RemoveContainer" containerID="8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.443299 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0\": container with ID starting with 8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0 not found: ID does not exist" containerID="8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.443351 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0"} err="failed to get container status \"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0\": rpc error: code = NotFound desc = could not find container \"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0\": container with ID starting with 8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.443383 4744 scope.go:117] "RemoveContainer" containerID="e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.443726 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\": container with ID starting with e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb not found: ID does not exist" containerID="e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.443831 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb"} err="failed to get container status \"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\": rpc error: code = NotFound desc = could not find container \"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\": container with ID starting with e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.443956 4744 scope.go:117] "RemoveContainer" containerID="1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.444512 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\": container with ID starting with 1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe not found: ID does not exist" containerID="1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.444542 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe"} err="failed to get container status \"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\": rpc error: code = NotFound desc = could not find container \"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\": container with ID starting with 1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.444592 4744 scope.go:117] "RemoveContainer" containerID="4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.444890 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\": container with ID starting with 4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64 not found: ID does not exist" containerID="4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.444916 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64"} err="failed to get container status \"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\": rpc error: code = NotFound desc = could not find container \"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\": container with ID starting with 4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.444952 4744 scope.go:117] "RemoveContainer" containerID="aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.445201 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\": container with ID starting with aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d not found: ID does not exist" containerID="aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.445222 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d"} err="failed to get container status \"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\": rpc error: code = NotFound desc = could not find container \"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\": container with ID starting with aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.445238 4744 scope.go:117] "RemoveContainer" containerID="c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.445464 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\": container with ID starting with c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c not found: ID does not exist" containerID="c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.445487 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c"} err="failed to get container status \"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\": rpc error: code = NotFound desc = could not find container \"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\": container with ID starting with c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.445508 4744 scope.go:117] "RemoveContainer" containerID="a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.445719 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\": container with ID starting with a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603 not found: ID does not exist" containerID="a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.445741 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603"} err="failed to get container status \"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\": rpc error: code = NotFound desc = could not find container \"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\": container with ID starting with a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.445757 4744 scope.go:117] "RemoveContainer" containerID="88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.445963 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\": container with ID starting with 88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558 not found: ID does not exist" containerID="88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.445987 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558"} err="failed to get container status \"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\": rpc error: code = NotFound desc = could not find container \"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\": container with ID starting with 88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.446004 4744 scope.go:117] "RemoveContainer" containerID="aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.446193 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\": container with ID starting with aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0 not found: ID does not exist" containerID="aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.446217 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0"} err="failed to get container status \"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\": rpc error: code = NotFound desc = could not find container \"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\": container with ID starting with aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.446233 4744 scope.go:117] "RemoveContainer" containerID="a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b" Oct 03 16:35:01 crc kubenswrapper[4744]: E1003 16:35:01.446493 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\": container with ID starting with a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b not found: ID does not exist" containerID="a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.446577 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b"} err="failed to get container status \"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\": rpc error: code = NotFound desc = could not find container \"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\": container with ID starting with a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.446606 4744 scope.go:117] "RemoveContainer" containerID="8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.446936 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0"} err="failed to get container status \"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0\": rpc error: code = NotFound desc = could not find container \"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0\": container with ID starting with 8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.446956 4744 scope.go:117] "RemoveContainer" containerID="e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.447285 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb"} err="failed to get container status \"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\": rpc error: code = NotFound desc = could not find container \"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\": container with ID starting with e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.447306 4744 scope.go:117] "RemoveContainer" containerID="1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.447675 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe"} err="failed to get container status \"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\": rpc error: code = NotFound desc = could not find container \"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\": container with ID starting with 1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.447701 4744 scope.go:117] "RemoveContainer" containerID="4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.447927 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64"} err="failed to get container status \"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\": rpc error: code = NotFound desc = could not find container \"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\": container with ID starting with 4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.447946 4744 scope.go:117] "RemoveContainer" containerID="aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.448277 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d"} err="failed to get container status \"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\": rpc error: code = NotFound desc = could not find container \"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\": container with ID starting with aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.448302 4744 scope.go:117] "RemoveContainer" containerID="c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.448503 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c"} err="failed to get container status \"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\": rpc error: code = NotFound desc = could not find container \"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\": container with ID starting with c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.448547 4744 scope.go:117] "RemoveContainer" containerID="a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.448758 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603"} err="failed to get container status \"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\": rpc error: code = NotFound desc = could not find container \"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\": container with ID starting with a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.448776 4744 scope.go:117] "RemoveContainer" containerID="88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.449111 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558"} err="failed to get container status \"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\": rpc error: code = NotFound desc = could not find container \"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\": container with ID starting with 88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.449129 4744 scope.go:117] "RemoveContainer" containerID="aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.449327 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0"} err="failed to get container status \"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\": rpc error: code = NotFound desc = could not find container \"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\": container with ID starting with aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.449345 4744 scope.go:117] "RemoveContainer" containerID="a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.449668 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b"} err="failed to get container status \"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\": rpc error: code = NotFound desc = could not find container \"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\": container with ID starting with a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.449688 4744 scope.go:117] "RemoveContainer" containerID="8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.449985 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0"} err="failed to get container status \"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0\": rpc error: code = NotFound desc = could not find container \"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0\": container with ID starting with 8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.450007 4744 scope.go:117] "RemoveContainer" containerID="e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.450287 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb"} err="failed to get container status \"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\": rpc error: code = NotFound desc = could not find container \"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\": container with ID starting with e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.450479 4744 scope.go:117] "RemoveContainer" containerID="1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.450910 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe"} err="failed to get container status \"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\": rpc error: code = NotFound desc = could not find container \"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\": container with ID starting with 1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.450936 4744 scope.go:117] "RemoveContainer" containerID="4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.451213 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64"} err="failed to get container status \"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\": rpc error: code = NotFound desc = could not find container \"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\": container with ID starting with 4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.451234 4744 scope.go:117] "RemoveContainer" containerID="aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.451498 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d"} err="failed to get container status \"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\": rpc error: code = NotFound desc = could not find container \"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\": container with ID starting with aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.451531 4744 scope.go:117] "RemoveContainer" containerID="c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.451841 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c"} err="failed to get container status \"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\": rpc error: code = NotFound desc = could not find container \"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\": container with ID starting with c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.451860 4744 scope.go:117] "RemoveContainer" containerID="a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.452053 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603"} err="failed to get container status \"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\": rpc error: code = NotFound desc = could not find container \"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\": container with ID starting with a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.452073 4744 scope.go:117] "RemoveContainer" containerID="88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.452348 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558"} err="failed to get container status \"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\": rpc error: code = NotFound desc = could not find container \"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\": container with ID starting with 88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.452368 4744 scope.go:117] "RemoveContainer" containerID="aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.452680 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0"} err="failed to get container status \"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\": rpc error: code = NotFound desc = could not find container \"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\": container with ID starting with aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.452703 4744 scope.go:117] "RemoveContainer" containerID="a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.454803 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b"} err="failed to get container status \"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\": rpc error: code = NotFound desc = could not find container \"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\": container with ID starting with a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.454823 4744 scope.go:117] "RemoveContainer" containerID="8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.455122 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0"} err="failed to get container status \"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0\": rpc error: code = NotFound desc = could not find container \"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0\": container with ID starting with 8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.455142 4744 scope.go:117] "RemoveContainer" containerID="e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.455481 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb"} err="failed to get container status \"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\": rpc error: code = NotFound desc = could not find container \"e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb\": container with ID starting with e2be290d94520cd2d33bcd0ec88bdbe7d49603a66f3608e02d7ce60309d243bb not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.455504 4744 scope.go:117] "RemoveContainer" containerID="1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.455802 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe"} err="failed to get container status \"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\": rpc error: code = NotFound desc = could not find container \"1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe\": container with ID starting with 1ac46256c44fbbb44d758adb776ee3eb2350244dbdab72a2f3bdd4ea231a49fe not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.455822 4744 scope.go:117] "RemoveContainer" containerID="4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.456025 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64"} err="failed to get container status \"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\": rpc error: code = NotFound desc = could not find container \"4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64\": container with ID starting with 4e52bccc9e7b91cb7ec346d0325715eb11a01dc23613df34caa7576a11932a64 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.456045 4744 scope.go:117] "RemoveContainer" containerID="aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.456661 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d"} err="failed to get container status \"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\": rpc error: code = NotFound desc = could not find container \"aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d\": container with ID starting with aff08fb902cc52b9df684e200baa3663f4de61521c1b4b11e16390292dca2a8d not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.456681 4744 scope.go:117] "RemoveContainer" containerID="c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.457000 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c"} err="failed to get container status \"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\": rpc error: code = NotFound desc = could not find container \"c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c\": container with ID starting with c6b950c733e8858c88b07b62f4796b7d1e106bb9cc0b5314160d7a489dba0b8c not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.457043 4744 scope.go:117] "RemoveContainer" containerID="a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.457351 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603"} err="failed to get container status \"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\": rpc error: code = NotFound desc = could not find container \"a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603\": container with ID starting with a346b882df1d3deefabf46f10a53339f4fef820b6070e0f89a931fd6b1cd1603 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.457372 4744 scope.go:117] "RemoveContainer" containerID="88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.457661 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558"} err="failed to get container status \"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\": rpc error: code = NotFound desc = could not find container \"88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558\": container with ID starting with 88d990e036eecef0cdc566bc559010b00c8c3c3adae46d37df007d7320ada558 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.457680 4744 scope.go:117] "RemoveContainer" containerID="aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.458142 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0"} err="failed to get container status \"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\": rpc error: code = NotFound desc = could not find container \"aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0\": container with ID starting with aa258bc66467c99410eb91fe47553844f1dee965965a1032378635b00d4c03d0 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.458162 4744 scope.go:117] "RemoveContainer" containerID="a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.458446 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b"} err="failed to get container status \"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\": rpc error: code = NotFound desc = could not find container \"a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b\": container with ID starting with a5ab94159686323028ba92f99977114e69e036214cfb45c342812bb4cb416a3b not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.458466 4744 scope.go:117] "RemoveContainer" containerID="8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.458704 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0"} err="failed to get container status \"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0\": rpc error: code = NotFound desc = could not find container \"8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0\": container with ID starting with 8fa6580a61d02d5966bc635f1114f2a898477e41878313ceef999dfc8991bcf0 not found: ID does not exist" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479349 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-ovnkube-script-lib\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479459 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-env-overrides\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479495 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-kubelet\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479511 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-slash\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479572 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-cni-bin\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479599 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-ovn-node-metrics-cert\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479624 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-run-ovn-kubernetes\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479646 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-run-systemd\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479648 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-kubelet\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479722 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-log-socket\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479671 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-log-socket\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479678 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-cni-bin\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479722 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-slash\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479772 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-run-systemd\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479799 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-run-ovn-kubernetes\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479940 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-cni-netd\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.479983 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-run-ovn\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.480016 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.480051 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-ovnkube-config\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.480115 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-run-netns\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.480181 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-systemd-units\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.480211 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk5s6\" (UniqueName: \"kubernetes.io/projected/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-kube-api-access-xk5s6\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.480371 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-cni-netd\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.480355 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.480628 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-systemd-units\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.480773 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-node-log\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.480791 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-host-run-netns\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.480827 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-run-ovn\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.480895 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-node-log\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.480974 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-run-openvswitch\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.481037 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-var-lib-openvswitch\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.481079 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-etc-openvswitch\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.483070 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.483128 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3436425f-b9cf-4302-b5ad-ccb92d597aed-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.481083 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-run-openvswitch\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.480987 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-env-overrides\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.483153 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxw7s\" (UniqueName: \"kubernetes.io/projected/3436425f-b9cf-4302-b5ad-ccb92d597aed-kube-api-access-nxw7s\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.481165 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-etc-openvswitch\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.481113 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-var-lib-openvswitch\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.481330 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-ovnkube-config\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.483314 4744 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.483398 4744 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3436425f-b9cf-4302-b5ad-ccb92d597aed-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.483774 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-ovn-node-metrics-cert\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.484189 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-ovnkube-script-lib\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.499462 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk5s6\" (UniqueName: \"kubernetes.io/projected/1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf-kube-api-access-xk5s6\") pod \"ovnkube-node-ppskd\" (UID: \"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf\") " pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.574187 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:01 crc kubenswrapper[4744]: W1003 16:35:01.595866 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e9cb891_b6dc_4ffe_ab7e_bff19cb02bcf.slice/crio-86bcd3eb0e266864cd1514da002798796552b02dc319d8bacc527bbcebca3a48 WatchSource:0}: Error finding container 86bcd3eb0e266864cd1514da002798796552b02dc319d8bacc527bbcebca3a48: Status 404 returned error can't find the container with id 86bcd3eb0e266864cd1514da002798796552b02dc319d8bacc527bbcebca3a48 Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.611587 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jw5rc"] Oct 03 16:35:01 crc kubenswrapper[4744]: I1003 16:35:01.617794 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jw5rc"] Oct 03 16:35:02 crc kubenswrapper[4744]: I1003 16:35:02.280936 4744 generic.go:334] "Generic (PLEG): container finished" podID="1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf" containerID="ef533156a91f64c8b3f203279ab78428171d8dd3205ddb317d99108497b0650d" exitCode=0 Oct 03 16:35:02 crc kubenswrapper[4744]: I1003 16:35:02.281023 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" event={"ID":"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf","Type":"ContainerDied","Data":"ef533156a91f64c8b3f203279ab78428171d8dd3205ddb317d99108497b0650d"} Oct 03 16:35:02 crc kubenswrapper[4744]: I1003 16:35:02.281488 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" event={"ID":"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf","Type":"ContainerStarted","Data":"86bcd3eb0e266864cd1514da002798796552b02dc319d8bacc527bbcebca3a48"} Oct 03 16:35:02 crc kubenswrapper[4744]: I1003 16:35:02.903927 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3436425f-b9cf-4302-b5ad-ccb92d597aed" path="/var/lib/kubelet/pods/3436425f-b9cf-4302-b5ad-ccb92d597aed/volumes" Oct 03 16:35:03 crc kubenswrapper[4744]: I1003 16:35:03.291317 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" event={"ID":"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf","Type":"ContainerStarted","Data":"0d35f4a404122434f8205b7fb76f500fa1e6e13c47ca8005e4d311d7420b8cab"} Oct 03 16:35:03 crc kubenswrapper[4744]: I1003 16:35:03.292551 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" event={"ID":"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf","Type":"ContainerStarted","Data":"1a67b0410df51169dae6ad1b8290b93b429c8d8503d3f2b6ec8e2d0e5637e222"} Oct 03 16:35:03 crc kubenswrapper[4744]: I1003 16:35:03.292847 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" event={"ID":"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf","Type":"ContainerStarted","Data":"f3d5421e9d40f7f51f4f7cb8ae62811e38204d7af138fa42f93579c8024cfe0f"} Oct 03 16:35:03 crc kubenswrapper[4744]: I1003 16:35:03.292857 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" event={"ID":"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf","Type":"ContainerStarted","Data":"737d7a6d759285c6ea1686f21be413f258fdc3585e16264de6c99c86617acdf8"} Oct 03 16:35:03 crc kubenswrapper[4744]: I1003 16:35:03.292866 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" event={"ID":"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf","Type":"ContainerStarted","Data":"13f039e4b81a420bad994d82cd644cc3334e42fd2e0faa1aba06e33199abe631"} Oct 03 16:35:03 crc kubenswrapper[4744]: I1003 16:35:03.292875 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" event={"ID":"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf","Type":"ContainerStarted","Data":"36270f9149a6c1db08488a1271a476d0115de16359b5825c7b416bbb7a754b31"} Oct 03 16:35:04 crc kubenswrapper[4744]: I1003 16:35:04.668803 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:35:04 crc kubenswrapper[4744]: I1003 16:35:04.668900 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:35:04 crc kubenswrapper[4744]: I1003 16:35:04.668963 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:35:04 crc kubenswrapper[4744]: I1003 16:35:04.669655 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c5619e90872156a7cc8093cc5e7ceeb12da127b0d8d653ccf7ad40c5326f1a97"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 16:35:04 crc kubenswrapper[4744]: I1003 16:35:04.669710 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://c5619e90872156a7cc8093cc5e7ceeb12da127b0d8d653ccf7ad40c5326f1a97" gracePeriod=600 Oct 03 16:35:05 crc kubenswrapper[4744]: I1003 16:35:05.308293 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" event={"ID":"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf","Type":"ContainerStarted","Data":"952027932cfa604cd0cbe2b10f5c1b1f3e5864d8e3a0c800917580cac47521ca"} Oct 03 16:35:05 crc kubenswrapper[4744]: I1003 16:35:05.311187 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="c5619e90872156a7cc8093cc5e7ceeb12da127b0d8d653ccf7ad40c5326f1a97" exitCode=0 Oct 03 16:35:05 crc kubenswrapper[4744]: I1003 16:35:05.311260 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"c5619e90872156a7cc8093cc5e7ceeb12da127b0d8d653ccf7ad40c5326f1a97"} Oct 03 16:35:05 crc kubenswrapper[4744]: I1003 16:35:05.311354 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"08b50c5cd69a2f26327ea9dc4a127dffa53b9765528a51178793482f7f49b17f"} Oct 03 16:35:05 crc kubenswrapper[4744]: I1003 16:35:05.311395 4744 scope.go:117] "RemoveContainer" containerID="ebd68629db9100ac91296997be91b152693b5453b3c435c930ae88e90650ca93" Oct 03 16:35:05 crc kubenswrapper[4744]: I1003 16:35:05.581426 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-xjmdh" Oct 03 16:35:08 crc kubenswrapper[4744]: I1003 16:35:08.343380 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" event={"ID":"1e9cb891-b6dc-4ffe-ab7e-bff19cb02bcf","Type":"ContainerStarted","Data":"14125db61e32afa096d97f621e6ebc65556bdb6c499864b1b73573077a79331c"} Oct 03 16:35:08 crc kubenswrapper[4744]: I1003 16:35:08.344241 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:08 crc kubenswrapper[4744]: I1003 16:35:08.344276 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:08 crc kubenswrapper[4744]: I1003 16:35:08.344289 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:08 crc kubenswrapper[4744]: I1003 16:35:08.402685 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" podStartSLOduration=7.402662918 podStartE2EDuration="7.402662918s" podCreationTimestamp="2025-10-03 16:35:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:35:08.393707249 +0000 UTC m=+634.673583145" watchObservedRunningTime="2025-10-03 16:35:08.402662918 +0000 UTC m=+634.682538814" Oct 03 16:35:08 crc kubenswrapper[4744]: I1003 16:35:08.411192 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:08 crc kubenswrapper[4744]: I1003 16:35:08.422033 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:12 crc kubenswrapper[4744]: I1003 16:35:12.892007 4744 scope.go:117] "RemoveContainer" containerID="b9031bf515dd80acd90501c31a1f575672257c7875c2e6b608cf36a4e51744a4" Oct 03 16:35:12 crc kubenswrapper[4744]: E1003 16:35:12.892795 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-vqz6q_openshift-multus(84192d79-366a-453f-b70b-aefa4537ec4c)\"" pod="openshift-multus/multus-vqz6q" podUID="84192d79-366a-453f-b70b-aefa4537ec4c" Oct 03 16:35:26 crc kubenswrapper[4744]: I1003 16:35:26.892260 4744 scope.go:117] "RemoveContainer" containerID="b9031bf515dd80acd90501c31a1f575672257c7875c2e6b608cf36a4e51744a4" Oct 03 16:35:27 crc kubenswrapper[4744]: I1003 16:35:27.479063 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vqz6q_84192d79-366a-453f-b70b-aefa4537ec4c/kube-multus/2.log" Oct 03 16:35:27 crc kubenswrapper[4744]: I1003 16:35:27.480462 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vqz6q_84192d79-366a-453f-b70b-aefa4537ec4c/kube-multus/1.log" Oct 03 16:35:27 crc kubenswrapper[4744]: I1003 16:35:27.480535 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vqz6q" event={"ID":"84192d79-366a-453f-b70b-aefa4537ec4c","Type":"ContainerStarted","Data":"88d547234a8e34766c120e81730dcd15888503b7b62da79f082352b612d942d9"} Oct 03 16:35:31 crc kubenswrapper[4744]: I1003 16:35:31.607485 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ppskd" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.268601 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph"] Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.270148 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.273577 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-cvbbb" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.273772 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.273915 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.359027 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log\" (UniqueName: \"kubernetes.io/empty-dir/af690e2a-fac7-4b2b-9ca0-4473e9f6989d-log\") pod \"ceph\" (UID: \"af690e2a-fac7-4b2b-9ca0-4473e9f6989d\") " pod="openstack/ceph" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.359101 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5s2q\" (UniqueName: \"kubernetes.io/projected/af690e2a-fac7-4b2b-9ca0-4473e9f6989d-kube-api-access-b5s2q\") pod \"ceph\" (UID: \"af690e2a-fac7-4b2b-9ca0-4473e9f6989d\") " pod="openstack/ceph" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.359247 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/af690e2a-fac7-4b2b-9ca0-4473e9f6989d-data\") pod \"ceph\" (UID: \"af690e2a-fac7-4b2b-9ca0-4473e9f6989d\") " pod="openstack/ceph" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.359581 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/empty-dir/af690e2a-fac7-4b2b-9ca0-4473e9f6989d-run\") pod \"ceph\" (UID: \"af690e2a-fac7-4b2b-9ca0-4473e9f6989d\") " pod="openstack/ceph" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.460997 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/empty-dir/af690e2a-fac7-4b2b-9ca0-4473e9f6989d-run\") pod \"ceph\" (UID: \"af690e2a-fac7-4b2b-9ca0-4473e9f6989d\") " pod="openstack/ceph" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.461085 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log\" (UniqueName: \"kubernetes.io/empty-dir/af690e2a-fac7-4b2b-9ca0-4473e9f6989d-log\") pod \"ceph\" (UID: \"af690e2a-fac7-4b2b-9ca0-4473e9f6989d\") " pod="openstack/ceph" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.461118 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5s2q\" (UniqueName: \"kubernetes.io/projected/af690e2a-fac7-4b2b-9ca0-4473e9f6989d-kube-api-access-b5s2q\") pod \"ceph\" (UID: \"af690e2a-fac7-4b2b-9ca0-4473e9f6989d\") " pod="openstack/ceph" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.461162 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/af690e2a-fac7-4b2b-9ca0-4473e9f6989d-data\") pod \"ceph\" (UID: \"af690e2a-fac7-4b2b-9ca0-4473e9f6989d\") " pod="openstack/ceph" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.461904 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/af690e2a-fac7-4b2b-9ca0-4473e9f6989d-data\") pod \"ceph\" (UID: \"af690e2a-fac7-4b2b-9ca0-4473e9f6989d\") " pod="openstack/ceph" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.462054 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log\" (UniqueName: \"kubernetes.io/empty-dir/af690e2a-fac7-4b2b-9ca0-4473e9f6989d-log\") pod \"ceph\" (UID: \"af690e2a-fac7-4b2b-9ca0-4473e9f6989d\") " pod="openstack/ceph" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.466089 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/empty-dir/af690e2a-fac7-4b2b-9ca0-4473e9f6989d-run\") pod \"ceph\" (UID: \"af690e2a-fac7-4b2b-9ca0-4473e9f6989d\") " pod="openstack/ceph" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.498100 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5s2q\" (UniqueName: \"kubernetes.io/projected/af690e2a-fac7-4b2b-9ca0-4473e9f6989d-kube-api-access-b5s2q\") pod \"ceph\" (UID: \"af690e2a-fac7-4b2b-9ca0-4473e9f6989d\") " pod="openstack/ceph" Oct 03 16:35:33 crc kubenswrapper[4744]: I1003 16:35:33.598907 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph" Oct 03 16:35:34 crc kubenswrapper[4744]: I1003 16:35:34.529810 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph" event={"ID":"af690e2a-fac7-4b2b-9ca0-4473e9f6989d","Type":"ContainerStarted","Data":"18025acc77e1484b64c3ad916679321ae6b7baf0dbdaecc6d510fbd7974c90c4"} Oct 03 16:35:35 crc kubenswrapper[4744]: I1003 16:35:35.075325 4744 scope.go:117] "RemoveContainer" containerID="d3888e7ce3b4e4e7967e3c7ec2e5ef8b7bee43a7742087ca6ea6f44a3d72e45c" Oct 03 16:35:35 crc kubenswrapper[4744]: I1003 16:35:35.538018 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vqz6q_84192d79-366a-453f-b70b-aefa4537ec4c/kube-multus/2.log" Oct 03 16:35:50 crc kubenswrapper[4744]: I1003 16:35:50.641551 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph" event={"ID":"af690e2a-fac7-4b2b-9ca0-4473e9f6989d","Type":"ContainerStarted","Data":"b079935023776d546d26dc04917be5916d60da213e88d031050dfbc898ac6107"} Oct 03 16:35:50 crc kubenswrapper[4744]: I1003 16:35:50.677138 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph" podStartSLOduration=1.089266857 podStartE2EDuration="17.677100673s" podCreationTimestamp="2025-10-03 16:35:33 +0000 UTC" firstStartedPulling="2025-10-03 16:35:33.629543481 +0000 UTC m=+659.909419377" lastFinishedPulling="2025-10-03 16:35:50.217377307 +0000 UTC m=+676.497253193" observedRunningTime="2025-10-03 16:35:50.660428671 +0000 UTC m=+676.940304597" watchObservedRunningTime="2025-10-03 16:35:50.677100673 +0000 UTC m=+676.956976589" Oct 03 16:37:04 crc kubenswrapper[4744]: I1003 16:37:04.669056 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:37:04 crc kubenswrapper[4744]: I1003 16:37:04.669767 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:37:13 crc kubenswrapper[4744]: I1003 16:37:13.771307 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xb6fx"] Oct 03 16:37:13 crc kubenswrapper[4744]: I1003 16:37:13.772573 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" podUID="1bd33bcb-22df-4c14-9ccd-8661ff4615d5" containerName="controller-manager" containerID="cri-o://4f37b00d6c7dc0c25e578e4f29ba45d009b18d77ffd56d78cae9f7dd5607a1bb" gracePeriod=30 Oct 03 16:37:13 crc kubenswrapper[4744]: I1003 16:37:13.867535 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k"] Oct 03 16:37:13 crc kubenswrapper[4744]: I1003 16:37:13.868090 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" podUID="b81e969d-bc65-4b6a-aaa3-ced704e3c744" containerName="route-controller-manager" containerID="cri-o://de3f5609b11d80898bda549e00f708e4abf6cfcce1093c61b92861639829103e" gracePeriod=30 Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.174220 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.180116 4744 generic.go:334] "Generic (PLEG): container finished" podID="1bd33bcb-22df-4c14-9ccd-8661ff4615d5" containerID="4f37b00d6c7dc0c25e578e4f29ba45d009b18d77ffd56d78cae9f7dd5607a1bb" exitCode=0 Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.180189 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.180201 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" event={"ID":"1bd33bcb-22df-4c14-9ccd-8661ff4615d5","Type":"ContainerDied","Data":"4f37b00d6c7dc0c25e578e4f29ba45d009b18d77ffd56d78cae9f7dd5607a1bb"} Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.180269 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xb6fx" event={"ID":"1bd33bcb-22df-4c14-9ccd-8661ff4615d5","Type":"ContainerDied","Data":"32337321d3e006f6082f53d1339d25aa997bffc54d8d206c784b2c8c293b8199"} Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.180301 4744 scope.go:117] "RemoveContainer" containerID="4f37b00d6c7dc0c25e578e4f29ba45d009b18d77ffd56d78cae9f7dd5607a1bb" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.183606 4744 generic.go:334] "Generic (PLEG): container finished" podID="b81e969d-bc65-4b6a-aaa3-ced704e3c744" containerID="de3f5609b11d80898bda549e00f708e4abf6cfcce1093c61b92861639829103e" exitCode=0 Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.183738 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" event={"ID":"b81e969d-bc65-4b6a-aaa3-ced704e3c744","Type":"ContainerDied","Data":"de3f5609b11d80898bda549e00f708e4abf6cfcce1093c61b92861639829103e"} Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.216090 4744 scope.go:117] "RemoveContainer" containerID="4f37b00d6c7dc0c25e578e4f29ba45d009b18d77ffd56d78cae9f7dd5607a1bb" Oct 03 16:37:14 crc kubenswrapper[4744]: E1003 16:37:14.219030 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f37b00d6c7dc0c25e578e4f29ba45d009b18d77ffd56d78cae9f7dd5607a1bb\": container with ID starting with 4f37b00d6c7dc0c25e578e4f29ba45d009b18d77ffd56d78cae9f7dd5607a1bb not found: ID does not exist" containerID="4f37b00d6c7dc0c25e578e4f29ba45d009b18d77ffd56d78cae9f7dd5607a1bb" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.219070 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f37b00d6c7dc0c25e578e4f29ba45d009b18d77ffd56d78cae9f7dd5607a1bb"} err="failed to get container status \"4f37b00d6c7dc0c25e578e4f29ba45d009b18d77ffd56d78cae9f7dd5607a1bb\": rpc error: code = NotFound desc = could not find container \"4f37b00d6c7dc0c25e578e4f29ba45d009b18d77ffd56d78cae9f7dd5607a1bb\": container with ID starting with 4f37b00d6c7dc0c25e578e4f29ba45d009b18d77ffd56d78cae9f7dd5607a1bb not found: ID does not exist" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.249282 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.318082 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b81e969d-bc65-4b6a-aaa3-ced704e3c744-client-ca\") pod \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.318140 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvqgt\" (UniqueName: \"kubernetes.io/projected/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-kube-api-access-kvqgt\") pod \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.318194 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b81e969d-bc65-4b6a-aaa3-ced704e3c744-config\") pod \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.319197 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b81e969d-bc65-4b6a-aaa3-ced704e3c744-config" (OuterVolumeSpecName: "config") pod "b81e969d-bc65-4b6a-aaa3-ced704e3c744" (UID: "b81e969d-bc65-4b6a-aaa3-ced704e3c744"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.319215 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b81e969d-bc65-4b6a-aaa3-ced704e3c744-client-ca" (OuterVolumeSpecName: "client-ca") pod "b81e969d-bc65-4b6a-aaa3-ced704e3c744" (UID: "b81e969d-bc65-4b6a-aaa3-ced704e3c744"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.319252 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-config\") pod \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.319377 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-serving-cert\") pod \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.319410 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b81e969d-bc65-4b6a-aaa3-ced704e3c744-serving-cert\") pod \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.319643 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-client-ca\") pod \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.319680 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-proxy-ca-bundles\") pod \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\" (UID: \"1bd33bcb-22df-4c14-9ccd-8661ff4615d5\") " Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.319792 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kbcn\" (UniqueName: \"kubernetes.io/projected/b81e969d-bc65-4b6a-aaa3-ced704e3c744-kube-api-access-7kbcn\") pod \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\" (UID: \"b81e969d-bc65-4b6a-aaa3-ced704e3c744\") " Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.320011 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-config" (OuterVolumeSpecName: "config") pod "1bd33bcb-22df-4c14-9ccd-8661ff4615d5" (UID: "1bd33bcb-22df-4c14-9ccd-8661ff4615d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.320244 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1bd33bcb-22df-4c14-9ccd-8661ff4615d5" (UID: "1bd33bcb-22df-4c14-9ccd-8661ff4615d5"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.320303 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-client-ca" (OuterVolumeSpecName: "client-ca") pod "1bd33bcb-22df-4c14-9ccd-8661ff4615d5" (UID: "1bd33bcb-22df-4c14-9ccd-8661ff4615d5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.320320 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b81e969d-bc65-4b6a-aaa3-ced704e3c744-client-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.320404 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b81e969d-bc65-4b6a-aaa3-ced704e3c744-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.320426 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.324420 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-kube-api-access-kvqgt" (OuterVolumeSpecName: "kube-api-access-kvqgt") pod "1bd33bcb-22df-4c14-9ccd-8661ff4615d5" (UID: "1bd33bcb-22df-4c14-9ccd-8661ff4615d5"). InnerVolumeSpecName "kube-api-access-kvqgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.324456 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b81e969d-bc65-4b6a-aaa3-ced704e3c744-kube-api-access-7kbcn" (OuterVolumeSpecName: "kube-api-access-7kbcn") pod "b81e969d-bc65-4b6a-aaa3-ced704e3c744" (UID: "b81e969d-bc65-4b6a-aaa3-ced704e3c744"). InnerVolumeSpecName "kube-api-access-7kbcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.324593 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b81e969d-bc65-4b6a-aaa3-ced704e3c744-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b81e969d-bc65-4b6a-aaa3-ced704e3c744" (UID: "b81e969d-bc65-4b6a-aaa3-ced704e3c744"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.324786 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bd33bcb-22df-4c14-9ccd-8661ff4615d5" (UID: "1bd33bcb-22df-4c14-9ccd-8661ff4615d5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.422169 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvqgt\" (UniqueName: \"kubernetes.io/projected/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-kube-api-access-kvqgt\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.422218 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b81e969d-bc65-4b6a-aaa3-ced704e3c744-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.422236 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.422250 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-client-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.422262 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1bd33bcb-22df-4c14-9ccd-8661ff4615d5-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.422273 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kbcn\" (UniqueName: \"kubernetes.io/projected/b81e969d-bc65-4b6a-aaa3-ced704e3c744-kube-api-access-7kbcn\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.509698 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xb6fx"] Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.517519 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xb6fx"] Oct 03 16:37:14 crc kubenswrapper[4744]: I1003 16:37:14.899093 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bd33bcb-22df-4c14-9ccd-8661ff4615d5" path="/var/lib/kubelet/pods/1bd33bcb-22df-4c14-9ccd-8661ff4615d5/volumes" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.190549 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" event={"ID":"b81e969d-bc65-4b6a-aaa3-ced704e3c744","Type":"ContainerDied","Data":"6c6b7c72dc428d1cc626f395f7191a3259f657fe9b6f94e03f1e858ce7638baa"} Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.191032 4744 scope.go:117] "RemoveContainer" containerID="de3f5609b11d80898bda549e00f708e4abf6cfcce1093c61b92861639829103e" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.191165 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.212674 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k"] Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.218904 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xj27k"] Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.435663 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-745469777d-nfvn9"] Oct 03 16:37:15 crc kubenswrapper[4744]: E1003 16:37:15.436168 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b81e969d-bc65-4b6a-aaa3-ced704e3c744" containerName="route-controller-manager" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.436195 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b81e969d-bc65-4b6a-aaa3-ced704e3c744" containerName="route-controller-manager" Oct 03 16:37:15 crc kubenswrapper[4744]: E1003 16:37:15.436232 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bd33bcb-22df-4c14-9ccd-8661ff4615d5" containerName="controller-manager" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.436242 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bd33bcb-22df-4c14-9ccd-8661ff4615d5" containerName="controller-manager" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.436391 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bd33bcb-22df-4c14-9ccd-8661ff4615d5" containerName="controller-manager" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.436409 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b81e969d-bc65-4b6a-aaa3-ced704e3c744" containerName="route-controller-manager" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.436920 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.438464 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z"] Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.439424 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.441087 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.442030 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.442812 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.442848 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.442819 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.442885 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.446367 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.446529 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.446697 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.447309 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.447667 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.447668 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.457360 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.457411 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z"] Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.462095 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-745469777d-nfvn9"] Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.540971 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a290f57-40a4-4deb-aec2-42777da19a66-client-ca\") pod \"route-controller-manager-746d44548c-x8x9z\" (UID: \"1a290f57-40a4-4deb-aec2-42777da19a66\") " pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.541053 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92a033e7-a697-49d6-ab9a-11d45ba38513-config\") pod \"controller-manager-745469777d-nfvn9\" (UID: \"92a033e7-a697-49d6-ab9a-11d45ba38513\") " pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.541152 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92a033e7-a697-49d6-ab9a-11d45ba38513-serving-cert\") pod \"controller-manager-745469777d-nfvn9\" (UID: \"92a033e7-a697-49d6-ab9a-11d45ba38513\") " pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.541377 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz4j6\" (UniqueName: \"kubernetes.io/projected/92a033e7-a697-49d6-ab9a-11d45ba38513-kube-api-access-zz4j6\") pod \"controller-manager-745469777d-nfvn9\" (UID: \"92a033e7-a697-49d6-ab9a-11d45ba38513\") " pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.541487 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a290f57-40a4-4deb-aec2-42777da19a66-config\") pod \"route-controller-manager-746d44548c-x8x9z\" (UID: \"1a290f57-40a4-4deb-aec2-42777da19a66\") " pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.541581 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4ght\" (UniqueName: \"kubernetes.io/projected/1a290f57-40a4-4deb-aec2-42777da19a66-kube-api-access-x4ght\") pod \"route-controller-manager-746d44548c-x8x9z\" (UID: \"1a290f57-40a4-4deb-aec2-42777da19a66\") " pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.541625 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a290f57-40a4-4deb-aec2-42777da19a66-serving-cert\") pod \"route-controller-manager-746d44548c-x8x9z\" (UID: \"1a290f57-40a4-4deb-aec2-42777da19a66\") " pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.541759 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92a033e7-a697-49d6-ab9a-11d45ba38513-client-ca\") pod \"controller-manager-745469777d-nfvn9\" (UID: \"92a033e7-a697-49d6-ab9a-11d45ba38513\") " pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.541950 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/92a033e7-a697-49d6-ab9a-11d45ba38513-proxy-ca-bundles\") pod \"controller-manager-745469777d-nfvn9\" (UID: \"92a033e7-a697-49d6-ab9a-11d45ba38513\") " pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.643029 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a290f57-40a4-4deb-aec2-42777da19a66-serving-cert\") pod \"route-controller-manager-746d44548c-x8x9z\" (UID: \"1a290f57-40a4-4deb-aec2-42777da19a66\") " pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.643095 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92a033e7-a697-49d6-ab9a-11d45ba38513-client-ca\") pod \"controller-manager-745469777d-nfvn9\" (UID: \"92a033e7-a697-49d6-ab9a-11d45ba38513\") " pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.643145 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/92a033e7-a697-49d6-ab9a-11d45ba38513-proxy-ca-bundles\") pod \"controller-manager-745469777d-nfvn9\" (UID: \"92a033e7-a697-49d6-ab9a-11d45ba38513\") " pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.643188 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a290f57-40a4-4deb-aec2-42777da19a66-client-ca\") pod \"route-controller-manager-746d44548c-x8x9z\" (UID: \"1a290f57-40a4-4deb-aec2-42777da19a66\") " pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.643229 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92a033e7-a697-49d6-ab9a-11d45ba38513-config\") pod \"controller-manager-745469777d-nfvn9\" (UID: \"92a033e7-a697-49d6-ab9a-11d45ba38513\") " pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.643259 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92a033e7-a697-49d6-ab9a-11d45ba38513-serving-cert\") pod \"controller-manager-745469777d-nfvn9\" (UID: \"92a033e7-a697-49d6-ab9a-11d45ba38513\") " pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.643289 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz4j6\" (UniqueName: \"kubernetes.io/projected/92a033e7-a697-49d6-ab9a-11d45ba38513-kube-api-access-zz4j6\") pod \"controller-manager-745469777d-nfvn9\" (UID: \"92a033e7-a697-49d6-ab9a-11d45ba38513\") " pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.643324 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a290f57-40a4-4deb-aec2-42777da19a66-config\") pod \"route-controller-manager-746d44548c-x8x9z\" (UID: \"1a290f57-40a4-4deb-aec2-42777da19a66\") " pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.645003 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92a033e7-a697-49d6-ab9a-11d45ba38513-client-ca\") pod \"controller-manager-745469777d-nfvn9\" (UID: \"92a033e7-a697-49d6-ab9a-11d45ba38513\") " pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.645578 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a290f57-40a4-4deb-aec2-42777da19a66-client-ca\") pod \"route-controller-manager-746d44548c-x8x9z\" (UID: \"1a290f57-40a4-4deb-aec2-42777da19a66\") " pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.645615 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a290f57-40a4-4deb-aec2-42777da19a66-config\") pod \"route-controller-manager-746d44548c-x8x9z\" (UID: \"1a290f57-40a4-4deb-aec2-42777da19a66\") " pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.645850 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/92a033e7-a697-49d6-ab9a-11d45ba38513-proxy-ca-bundles\") pod \"controller-manager-745469777d-nfvn9\" (UID: \"92a033e7-a697-49d6-ab9a-11d45ba38513\") " pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.646061 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4ght\" (UniqueName: \"kubernetes.io/projected/1a290f57-40a4-4deb-aec2-42777da19a66-kube-api-access-x4ght\") pod \"route-controller-manager-746d44548c-x8x9z\" (UID: \"1a290f57-40a4-4deb-aec2-42777da19a66\") " pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.649171 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92a033e7-a697-49d6-ab9a-11d45ba38513-config\") pod \"controller-manager-745469777d-nfvn9\" (UID: \"92a033e7-a697-49d6-ab9a-11d45ba38513\") " pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.649627 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a290f57-40a4-4deb-aec2-42777da19a66-serving-cert\") pod \"route-controller-manager-746d44548c-x8x9z\" (UID: \"1a290f57-40a4-4deb-aec2-42777da19a66\") " pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.650048 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92a033e7-a697-49d6-ab9a-11d45ba38513-serving-cert\") pod \"controller-manager-745469777d-nfvn9\" (UID: \"92a033e7-a697-49d6-ab9a-11d45ba38513\") " pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.660554 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz4j6\" (UniqueName: \"kubernetes.io/projected/92a033e7-a697-49d6-ab9a-11d45ba38513-kube-api-access-zz4j6\") pod \"controller-manager-745469777d-nfvn9\" (UID: \"92a033e7-a697-49d6-ab9a-11d45ba38513\") " pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.668174 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4ght\" (UniqueName: \"kubernetes.io/projected/1a290f57-40a4-4deb-aec2-42777da19a66-kube-api-access-x4ght\") pod \"route-controller-manager-746d44548c-x8x9z\" (UID: \"1a290f57-40a4-4deb-aec2-42777da19a66\") " pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.758810 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.766800 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.853374 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd"] Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.854742 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.857298 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.868981 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd"] Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.952775 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6a7d3bdf-d461-4fcb-810c-421130e8d524-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd\" (UID: \"6a7d3bdf-d461-4fcb-810c-421130e8d524\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.952827 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6a7d3bdf-d461-4fcb-810c-421130e8d524-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd\" (UID: \"6a7d3bdf-d461-4fcb-810c-421130e8d524\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" Oct 03 16:37:15 crc kubenswrapper[4744]: I1003 16:37:15.952861 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfnmn\" (UniqueName: \"kubernetes.io/projected/6a7d3bdf-d461-4fcb-810c-421130e8d524-kube-api-access-wfnmn\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd\" (UID: \"6a7d3bdf-d461-4fcb-810c-421130e8d524\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" Oct 03 16:37:16 crc kubenswrapper[4744]: I1003 16:37:16.003247 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z"] Oct 03 16:37:16 crc kubenswrapper[4744]: I1003 16:37:16.042578 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-745469777d-nfvn9"] Oct 03 16:37:16 crc kubenswrapper[4744]: I1003 16:37:16.054103 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6a7d3bdf-d461-4fcb-810c-421130e8d524-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd\" (UID: \"6a7d3bdf-d461-4fcb-810c-421130e8d524\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" Oct 03 16:37:16 crc kubenswrapper[4744]: I1003 16:37:16.054156 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6a7d3bdf-d461-4fcb-810c-421130e8d524-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd\" (UID: \"6a7d3bdf-d461-4fcb-810c-421130e8d524\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" Oct 03 16:37:16 crc kubenswrapper[4744]: I1003 16:37:16.054187 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfnmn\" (UniqueName: \"kubernetes.io/projected/6a7d3bdf-d461-4fcb-810c-421130e8d524-kube-api-access-wfnmn\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd\" (UID: \"6a7d3bdf-d461-4fcb-810c-421130e8d524\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" Oct 03 16:37:16 crc kubenswrapper[4744]: I1003 16:37:16.055390 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6a7d3bdf-d461-4fcb-810c-421130e8d524-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd\" (UID: \"6a7d3bdf-d461-4fcb-810c-421130e8d524\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" Oct 03 16:37:16 crc kubenswrapper[4744]: I1003 16:37:16.055621 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6a7d3bdf-d461-4fcb-810c-421130e8d524-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd\" (UID: \"6a7d3bdf-d461-4fcb-810c-421130e8d524\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" Oct 03 16:37:16 crc kubenswrapper[4744]: I1003 16:37:16.076791 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfnmn\" (UniqueName: \"kubernetes.io/projected/6a7d3bdf-d461-4fcb-810c-421130e8d524-kube-api-access-wfnmn\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd\" (UID: \"6a7d3bdf-d461-4fcb-810c-421130e8d524\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" Oct 03 16:37:16 crc kubenswrapper[4744]: I1003 16:37:16.174099 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" Oct 03 16:37:16 crc kubenswrapper[4744]: I1003 16:37:16.202650 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" event={"ID":"1a290f57-40a4-4deb-aec2-42777da19a66","Type":"ContainerStarted","Data":"63ec6411735670c5c221a49e51d140a732807e7b2a7a596d42c74cb181c2d614"} Oct 03 16:37:16 crc kubenswrapper[4744]: I1003 16:37:16.203958 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" event={"ID":"92a033e7-a697-49d6-ab9a-11d45ba38513","Type":"ContainerStarted","Data":"391a8df9fd0a9a3eca68bd17ba02e321d78d20bfc8b1958e2426ace84d6a0e41"} Oct 03 16:37:16 crc kubenswrapper[4744]: I1003 16:37:16.611981 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd"] Oct 03 16:37:16 crc kubenswrapper[4744]: I1003 16:37:16.901855 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b81e969d-bc65-4b6a-aaa3-ced704e3c744" path="/var/lib/kubelet/pods/b81e969d-bc65-4b6a-aaa3-ced704e3c744/volumes" Oct 03 16:37:17 crc kubenswrapper[4744]: I1003 16:37:17.216672 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" event={"ID":"92a033e7-a697-49d6-ab9a-11d45ba38513","Type":"ContainerStarted","Data":"98a24f53e060892181093124fa4594c54612878cf1d433a7d3d30df06a50484b"} Oct 03 16:37:17 crc kubenswrapper[4744]: I1003 16:37:17.218581 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:17 crc kubenswrapper[4744]: I1003 16:37:17.228199 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" Oct 03 16:37:17 crc kubenswrapper[4744]: I1003 16:37:17.240613 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" event={"ID":"1a290f57-40a4-4deb-aec2-42777da19a66","Type":"ContainerStarted","Data":"62c865ee648372229eacd69632b456bd5397c3a7a92b54462bd948bd1462cac4"} Oct 03 16:37:17 crc kubenswrapper[4744]: I1003 16:37:17.242518 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:17 crc kubenswrapper[4744]: I1003 16:37:17.242815 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-745469777d-nfvn9" podStartSLOduration=4.24278437 podStartE2EDuration="4.24278437s" podCreationTimestamp="2025-10-03 16:37:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:37:17.23968225 +0000 UTC m=+763.519558146" watchObservedRunningTime="2025-10-03 16:37:17.24278437 +0000 UTC m=+763.522660256" Oct 03 16:37:17 crc kubenswrapper[4744]: I1003 16:37:17.251356 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" Oct 03 16:37:17 crc kubenswrapper[4744]: I1003 16:37:17.255244 4744 generic.go:334] "Generic (PLEG): container finished" podID="6a7d3bdf-d461-4fcb-810c-421130e8d524" containerID="f8660f19c8d18e84d2fbc62620fcd4a13d1da55ecab7c541d10ddd74007b1f33" exitCode=0 Oct 03 16:37:17 crc kubenswrapper[4744]: I1003 16:37:17.256256 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" event={"ID":"6a7d3bdf-d461-4fcb-810c-421130e8d524","Type":"ContainerDied","Data":"f8660f19c8d18e84d2fbc62620fcd4a13d1da55ecab7c541d10ddd74007b1f33"} Oct 03 16:37:17 crc kubenswrapper[4744]: I1003 16:37:17.256330 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" event={"ID":"6a7d3bdf-d461-4fcb-810c-421130e8d524","Type":"ContainerStarted","Data":"c98d4f71c2f739ebeed5df04088db24bb491211551610b902a0b802ecb39a655"} Oct 03 16:37:17 crc kubenswrapper[4744]: I1003 16:37:17.317256 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-746d44548c-x8x9z" podStartSLOduration=4.317227238 podStartE2EDuration="4.317227238s" podCreationTimestamp="2025-10-03 16:37:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:37:17.311912621 +0000 UTC m=+763.591788517" watchObservedRunningTime="2025-10-03 16:37:17.317227238 +0000 UTC m=+763.597103154" Oct 03 16:37:18 crc kubenswrapper[4744]: I1003 16:37:18.200124 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cfwtm"] Oct 03 16:37:18 crc kubenswrapper[4744]: I1003 16:37:18.201966 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:18 crc kubenswrapper[4744]: I1003 16:37:18.204093 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cfwtm"] Oct 03 16:37:18 crc kubenswrapper[4744]: I1003 16:37:18.295332 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5b5646a-60f0-486c-8949-35b2c6965c9d-utilities\") pod \"redhat-operators-cfwtm\" (UID: \"e5b5646a-60f0-486c-8949-35b2c6965c9d\") " pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:18 crc kubenswrapper[4744]: I1003 16:37:18.295421 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx5mv\" (UniqueName: \"kubernetes.io/projected/e5b5646a-60f0-486c-8949-35b2c6965c9d-kube-api-access-jx5mv\") pod \"redhat-operators-cfwtm\" (UID: \"e5b5646a-60f0-486c-8949-35b2c6965c9d\") " pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:18 crc kubenswrapper[4744]: I1003 16:37:18.295464 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5b5646a-60f0-486c-8949-35b2c6965c9d-catalog-content\") pod \"redhat-operators-cfwtm\" (UID: \"e5b5646a-60f0-486c-8949-35b2c6965c9d\") " pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:18 crc kubenswrapper[4744]: I1003 16:37:18.397226 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5b5646a-60f0-486c-8949-35b2c6965c9d-utilities\") pod \"redhat-operators-cfwtm\" (UID: \"e5b5646a-60f0-486c-8949-35b2c6965c9d\") " pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:18 crc kubenswrapper[4744]: I1003 16:37:18.397298 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx5mv\" (UniqueName: \"kubernetes.io/projected/e5b5646a-60f0-486c-8949-35b2c6965c9d-kube-api-access-jx5mv\") pod \"redhat-operators-cfwtm\" (UID: \"e5b5646a-60f0-486c-8949-35b2c6965c9d\") " pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:18 crc kubenswrapper[4744]: I1003 16:37:18.397337 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5b5646a-60f0-486c-8949-35b2c6965c9d-catalog-content\") pod \"redhat-operators-cfwtm\" (UID: \"e5b5646a-60f0-486c-8949-35b2c6965c9d\") " pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:18 crc kubenswrapper[4744]: I1003 16:37:18.398296 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5b5646a-60f0-486c-8949-35b2c6965c9d-utilities\") pod \"redhat-operators-cfwtm\" (UID: \"e5b5646a-60f0-486c-8949-35b2c6965c9d\") " pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:18 crc kubenswrapper[4744]: I1003 16:37:18.400083 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5b5646a-60f0-486c-8949-35b2c6965c9d-catalog-content\") pod \"redhat-operators-cfwtm\" (UID: \"e5b5646a-60f0-486c-8949-35b2c6965c9d\") " pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:18 crc kubenswrapper[4744]: I1003 16:37:18.432597 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx5mv\" (UniqueName: \"kubernetes.io/projected/e5b5646a-60f0-486c-8949-35b2c6965c9d-kube-api-access-jx5mv\") pod \"redhat-operators-cfwtm\" (UID: \"e5b5646a-60f0-486c-8949-35b2c6965c9d\") " pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:18 crc kubenswrapper[4744]: I1003 16:37:18.531550 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:18 crc kubenswrapper[4744]: I1003 16:37:18.957377 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cfwtm"] Oct 03 16:37:18 crc kubenswrapper[4744]: W1003 16:37:18.965229 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5b5646a_60f0_486c_8949_35b2c6965c9d.slice/crio-20cba7b040b4bb16407dd7d3b7ba96674d5622da429097d9d1d835e431ce7c24 WatchSource:0}: Error finding container 20cba7b040b4bb16407dd7d3b7ba96674d5622da429097d9d1d835e431ce7c24: Status 404 returned error can't find the container with id 20cba7b040b4bb16407dd7d3b7ba96674d5622da429097d9d1d835e431ce7c24 Oct 03 16:37:19 crc kubenswrapper[4744]: I1003 16:37:19.275349 4744 generic.go:334] "Generic (PLEG): container finished" podID="6a7d3bdf-d461-4fcb-810c-421130e8d524" containerID="dcb43c8f2773443f026c64600607570d787b80371cf478560693113424472534" exitCode=0 Oct 03 16:37:19 crc kubenswrapper[4744]: I1003 16:37:19.275556 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" event={"ID":"6a7d3bdf-d461-4fcb-810c-421130e8d524","Type":"ContainerDied","Data":"dcb43c8f2773443f026c64600607570d787b80371cf478560693113424472534"} Oct 03 16:37:19 crc kubenswrapper[4744]: I1003 16:37:19.278653 4744 generic.go:334] "Generic (PLEG): container finished" podID="e5b5646a-60f0-486c-8949-35b2c6965c9d" containerID="910c6a57d2590f58fb6f6726d5a5486096cb43b934c836d3f9eaed34fcd41924" exitCode=0 Oct 03 16:37:19 crc kubenswrapper[4744]: I1003 16:37:19.278759 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfwtm" event={"ID":"e5b5646a-60f0-486c-8949-35b2c6965c9d","Type":"ContainerDied","Data":"910c6a57d2590f58fb6f6726d5a5486096cb43b934c836d3f9eaed34fcd41924"} Oct 03 16:37:19 crc kubenswrapper[4744]: I1003 16:37:19.278832 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfwtm" event={"ID":"e5b5646a-60f0-486c-8949-35b2c6965c9d","Type":"ContainerStarted","Data":"20cba7b040b4bb16407dd7d3b7ba96674d5622da429097d9d1d835e431ce7c24"} Oct 03 16:37:20 crc kubenswrapper[4744]: I1003 16:37:20.288832 4744 generic.go:334] "Generic (PLEG): container finished" podID="6a7d3bdf-d461-4fcb-810c-421130e8d524" containerID="144f24ab724e1431306592b4ae2095ecd9010e34337cd59cddd947030635dd72" exitCode=0 Oct 03 16:37:20 crc kubenswrapper[4744]: I1003 16:37:20.288897 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" event={"ID":"6a7d3bdf-d461-4fcb-810c-421130e8d524","Type":"ContainerDied","Data":"144f24ab724e1431306592b4ae2095ecd9010e34337cd59cddd947030635dd72"} Oct 03 16:37:21 crc kubenswrapper[4744]: I1003 16:37:21.298932 4744 generic.go:334] "Generic (PLEG): container finished" podID="e5b5646a-60f0-486c-8949-35b2c6965c9d" containerID="a857c1f14850cf7624323e8a173194275a7c5f2d7804fd5d3ef63bb425c612d8" exitCode=0 Oct 03 16:37:21 crc kubenswrapper[4744]: I1003 16:37:21.300485 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfwtm" event={"ID":"e5b5646a-60f0-486c-8949-35b2c6965c9d","Type":"ContainerDied","Data":"a857c1f14850cf7624323e8a173194275a7c5f2d7804fd5d3ef63bb425c612d8"} Oct 03 16:37:21 crc kubenswrapper[4744]: I1003 16:37:21.320203 4744 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 03 16:37:21 crc kubenswrapper[4744]: I1003 16:37:21.681727 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" Oct 03 16:37:21 crc kubenswrapper[4744]: I1003 16:37:21.750532 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6a7d3bdf-d461-4fcb-810c-421130e8d524-bundle\") pod \"6a7d3bdf-d461-4fcb-810c-421130e8d524\" (UID: \"6a7d3bdf-d461-4fcb-810c-421130e8d524\") " Oct 03 16:37:21 crc kubenswrapper[4744]: I1003 16:37:21.750643 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6a7d3bdf-d461-4fcb-810c-421130e8d524-util\") pod \"6a7d3bdf-d461-4fcb-810c-421130e8d524\" (UID: \"6a7d3bdf-d461-4fcb-810c-421130e8d524\") " Oct 03 16:37:21 crc kubenswrapper[4744]: I1003 16:37:21.750703 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfnmn\" (UniqueName: \"kubernetes.io/projected/6a7d3bdf-d461-4fcb-810c-421130e8d524-kube-api-access-wfnmn\") pod \"6a7d3bdf-d461-4fcb-810c-421130e8d524\" (UID: \"6a7d3bdf-d461-4fcb-810c-421130e8d524\") " Oct 03 16:37:21 crc kubenswrapper[4744]: I1003 16:37:21.752406 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a7d3bdf-d461-4fcb-810c-421130e8d524-bundle" (OuterVolumeSpecName: "bundle") pod "6a7d3bdf-d461-4fcb-810c-421130e8d524" (UID: "6a7d3bdf-d461-4fcb-810c-421130e8d524"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:37:21 crc kubenswrapper[4744]: I1003 16:37:21.759006 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a7d3bdf-d461-4fcb-810c-421130e8d524-kube-api-access-wfnmn" (OuterVolumeSpecName: "kube-api-access-wfnmn") pod "6a7d3bdf-d461-4fcb-810c-421130e8d524" (UID: "6a7d3bdf-d461-4fcb-810c-421130e8d524"). InnerVolumeSpecName "kube-api-access-wfnmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:37:21 crc kubenswrapper[4744]: I1003 16:37:21.764937 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a7d3bdf-d461-4fcb-810c-421130e8d524-util" (OuterVolumeSpecName: "util") pod "6a7d3bdf-d461-4fcb-810c-421130e8d524" (UID: "6a7d3bdf-d461-4fcb-810c-421130e8d524"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:37:21 crc kubenswrapper[4744]: I1003 16:37:21.852864 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6a7d3bdf-d461-4fcb-810c-421130e8d524-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:21 crc kubenswrapper[4744]: I1003 16:37:21.852914 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6a7d3bdf-d461-4fcb-810c-421130e8d524-util\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:21 crc kubenswrapper[4744]: I1003 16:37:21.852926 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfnmn\" (UniqueName: \"kubernetes.io/projected/6a7d3bdf-d461-4fcb-810c-421130e8d524-kube-api-access-wfnmn\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:22 crc kubenswrapper[4744]: I1003 16:37:22.313120 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" event={"ID":"6a7d3bdf-d461-4fcb-810c-421130e8d524","Type":"ContainerDied","Data":"c98d4f71c2f739ebeed5df04088db24bb491211551610b902a0b802ecb39a655"} Oct 03 16:37:22 crc kubenswrapper[4744]: I1003 16:37:22.314680 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c98d4f71c2f739ebeed5df04088db24bb491211551610b902a0b802ecb39a655" Oct 03 16:37:22 crc kubenswrapper[4744]: I1003 16:37:22.313217 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd" Oct 03 16:37:22 crc kubenswrapper[4744]: I1003 16:37:22.316438 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfwtm" event={"ID":"e5b5646a-60f0-486c-8949-35b2c6965c9d","Type":"ContainerStarted","Data":"e9a80f926cfc432ff65fd034cb3165edb19080d65e1488767a8a90196abce6f3"} Oct 03 16:37:22 crc kubenswrapper[4744]: I1003 16:37:22.340208 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cfwtm" podStartSLOduration=1.875173695 podStartE2EDuration="4.340189668s" podCreationTimestamp="2025-10-03 16:37:18 +0000 UTC" firstStartedPulling="2025-10-03 16:37:19.279988192 +0000 UTC m=+765.559864088" lastFinishedPulling="2025-10-03 16:37:21.745004165 +0000 UTC m=+768.024880061" observedRunningTime="2025-10-03 16:37:22.33755299 +0000 UTC m=+768.617428896" watchObservedRunningTime="2025-10-03 16:37:22.340189668 +0000 UTC m=+768.620065574" Oct 03 16:37:25 crc kubenswrapper[4744]: I1003 16:37:25.969328 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-qd582"] Oct 03 16:37:25 crc kubenswrapper[4744]: E1003 16:37:25.970082 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a7d3bdf-d461-4fcb-810c-421130e8d524" containerName="extract" Oct 03 16:37:25 crc kubenswrapper[4744]: I1003 16:37:25.970097 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a7d3bdf-d461-4fcb-810c-421130e8d524" containerName="extract" Oct 03 16:37:25 crc kubenswrapper[4744]: E1003 16:37:25.970111 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a7d3bdf-d461-4fcb-810c-421130e8d524" containerName="pull" Oct 03 16:37:25 crc kubenswrapper[4744]: I1003 16:37:25.970117 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a7d3bdf-d461-4fcb-810c-421130e8d524" containerName="pull" Oct 03 16:37:25 crc kubenswrapper[4744]: E1003 16:37:25.970125 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a7d3bdf-d461-4fcb-810c-421130e8d524" containerName="util" Oct 03 16:37:25 crc kubenswrapper[4744]: I1003 16:37:25.970131 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a7d3bdf-d461-4fcb-810c-421130e8d524" containerName="util" Oct 03 16:37:25 crc kubenswrapper[4744]: I1003 16:37:25.970232 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a7d3bdf-d461-4fcb-810c-421130e8d524" containerName="extract" Oct 03 16:37:25 crc kubenswrapper[4744]: I1003 16:37:25.970705 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-qd582" Oct 03 16:37:25 crc kubenswrapper[4744]: I1003 16:37:25.972951 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 03 16:37:25 crc kubenswrapper[4744]: I1003 16:37:25.973330 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 03 16:37:25 crc kubenswrapper[4744]: I1003 16:37:25.974031 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-wnm9h" Oct 03 16:37:25 crc kubenswrapper[4744]: I1003 16:37:25.984485 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-qd582"] Oct 03 16:37:26 crc kubenswrapper[4744]: I1003 16:37:26.110333 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdwqb\" (UniqueName: \"kubernetes.io/projected/af44daee-90fc-4ccd-a4dd-ae2513f0fe2c-kube-api-access-wdwqb\") pod \"nmstate-operator-858ddd8f98-qd582\" (UID: \"af44daee-90fc-4ccd-a4dd-ae2513f0fe2c\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-qd582" Oct 03 16:37:26 crc kubenswrapper[4744]: I1003 16:37:26.211587 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdwqb\" (UniqueName: \"kubernetes.io/projected/af44daee-90fc-4ccd-a4dd-ae2513f0fe2c-kube-api-access-wdwqb\") pod \"nmstate-operator-858ddd8f98-qd582\" (UID: \"af44daee-90fc-4ccd-a4dd-ae2513f0fe2c\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-qd582" Oct 03 16:37:26 crc kubenswrapper[4744]: I1003 16:37:26.231927 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdwqb\" (UniqueName: \"kubernetes.io/projected/af44daee-90fc-4ccd-a4dd-ae2513f0fe2c-kube-api-access-wdwqb\") pod \"nmstate-operator-858ddd8f98-qd582\" (UID: \"af44daee-90fc-4ccd-a4dd-ae2513f0fe2c\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-qd582" Oct 03 16:37:26 crc kubenswrapper[4744]: I1003 16:37:26.287323 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-qd582" Oct 03 16:37:26 crc kubenswrapper[4744]: I1003 16:37:26.763271 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-qd582"] Oct 03 16:37:27 crc kubenswrapper[4744]: I1003 16:37:27.349547 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-qd582" event={"ID":"af44daee-90fc-4ccd-a4dd-ae2513f0fe2c","Type":"ContainerStarted","Data":"4b9440616f69df0f69bf1cf4196cdca5a373852dd4630712f5df5f145b5e2eda"} Oct 03 16:37:28 crc kubenswrapper[4744]: I1003 16:37:28.532419 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:28 crc kubenswrapper[4744]: I1003 16:37:28.532934 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:28 crc kubenswrapper[4744]: I1003 16:37:28.582775 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:28 crc kubenswrapper[4744]: I1003 16:37:28.593858 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-49tm4"] Oct 03 16:37:28 crc kubenswrapper[4744]: I1003 16:37:28.595210 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:28 crc kubenswrapper[4744]: I1003 16:37:28.606763 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-49tm4"] Oct 03 16:37:28 crc kubenswrapper[4744]: I1003 16:37:28.748060 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc8695c-fc62-4788-9618-c0bed94eebd6-catalog-content\") pod \"community-operators-49tm4\" (UID: \"afc8695c-fc62-4788-9618-c0bed94eebd6\") " pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:28 crc kubenswrapper[4744]: I1003 16:37:28.748181 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l7hz\" (UniqueName: \"kubernetes.io/projected/afc8695c-fc62-4788-9618-c0bed94eebd6-kube-api-access-8l7hz\") pod \"community-operators-49tm4\" (UID: \"afc8695c-fc62-4788-9618-c0bed94eebd6\") " pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:28 crc kubenswrapper[4744]: I1003 16:37:28.748562 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc8695c-fc62-4788-9618-c0bed94eebd6-utilities\") pod \"community-operators-49tm4\" (UID: \"afc8695c-fc62-4788-9618-c0bed94eebd6\") " pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:28 crc kubenswrapper[4744]: I1003 16:37:28.850196 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc8695c-fc62-4788-9618-c0bed94eebd6-utilities\") pod \"community-operators-49tm4\" (UID: \"afc8695c-fc62-4788-9618-c0bed94eebd6\") " pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:28 crc kubenswrapper[4744]: I1003 16:37:28.850259 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc8695c-fc62-4788-9618-c0bed94eebd6-catalog-content\") pod \"community-operators-49tm4\" (UID: \"afc8695c-fc62-4788-9618-c0bed94eebd6\") " pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:28 crc kubenswrapper[4744]: I1003 16:37:28.850292 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l7hz\" (UniqueName: \"kubernetes.io/projected/afc8695c-fc62-4788-9618-c0bed94eebd6-kube-api-access-8l7hz\") pod \"community-operators-49tm4\" (UID: \"afc8695c-fc62-4788-9618-c0bed94eebd6\") " pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:28 crc kubenswrapper[4744]: I1003 16:37:28.917557 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc8695c-fc62-4788-9618-c0bed94eebd6-catalog-content\") pod \"community-operators-49tm4\" (UID: \"afc8695c-fc62-4788-9618-c0bed94eebd6\") " pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:28 crc kubenswrapper[4744]: I1003 16:37:28.917947 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc8695c-fc62-4788-9618-c0bed94eebd6-utilities\") pod \"community-operators-49tm4\" (UID: \"afc8695c-fc62-4788-9618-c0bed94eebd6\") " pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:28 crc kubenswrapper[4744]: I1003 16:37:28.933458 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l7hz\" (UniqueName: \"kubernetes.io/projected/afc8695c-fc62-4788-9618-c0bed94eebd6-kube-api-access-8l7hz\") pod \"community-operators-49tm4\" (UID: \"afc8695c-fc62-4788-9618-c0bed94eebd6\") " pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:29 crc kubenswrapper[4744]: I1003 16:37:29.226009 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:29 crc kubenswrapper[4744]: I1003 16:37:29.408274 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:29 crc kubenswrapper[4744]: I1003 16:37:29.899894 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-49tm4"] Oct 03 16:37:30 crc kubenswrapper[4744]: I1003 16:37:30.374435 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-qd582" event={"ID":"af44daee-90fc-4ccd-a4dd-ae2513f0fe2c","Type":"ContainerStarted","Data":"92592acf38c417f692d84af569acdbc5a407e26e21fbad83513fef1f5382fb2d"} Oct 03 16:37:30 crc kubenswrapper[4744]: I1003 16:37:30.376874 4744 generic.go:334] "Generic (PLEG): container finished" podID="afc8695c-fc62-4788-9618-c0bed94eebd6" containerID="937eb220f563d0f6f9dfa1e1b0b7324a4c48241f3bcabc6fa993e330a6d7f78d" exitCode=0 Oct 03 16:37:30 crc kubenswrapper[4744]: I1003 16:37:30.376937 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49tm4" event={"ID":"afc8695c-fc62-4788-9618-c0bed94eebd6","Type":"ContainerDied","Data":"937eb220f563d0f6f9dfa1e1b0b7324a4c48241f3bcabc6fa993e330a6d7f78d"} Oct 03 16:37:30 crc kubenswrapper[4744]: I1003 16:37:30.377078 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49tm4" event={"ID":"afc8695c-fc62-4788-9618-c0bed94eebd6","Type":"ContainerStarted","Data":"056b2eb320c5e152472bdda8b46b48f0a929303011708b50b20929dedbc8dd30"} Oct 03 16:37:30 crc kubenswrapper[4744]: I1003 16:37:30.407139 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-qd582" podStartSLOduration=2.683788269 podStartE2EDuration="5.407122217s" podCreationTimestamp="2025-10-03 16:37:25 +0000 UTC" firstStartedPulling="2025-10-03 16:37:26.790863785 +0000 UTC m=+773.070739681" lastFinishedPulling="2025-10-03 16:37:29.514197733 +0000 UTC m=+775.794073629" observedRunningTime="2025-10-03 16:37:30.404793727 +0000 UTC m=+776.684669663" watchObservedRunningTime="2025-10-03 16:37:30.407122217 +0000 UTC m=+776.686998123" Oct 03 16:37:31 crc kubenswrapper[4744]: I1003 16:37:31.385111 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49tm4" event={"ID":"afc8695c-fc62-4788-9618-c0bed94eebd6","Type":"ContainerStarted","Data":"0c4a0c15f78bf7c16af9edf410bca8e20028877da4bd9f64f3f6fbea0bc65264"} Oct 03 16:37:31 crc kubenswrapper[4744]: I1003 16:37:31.983972 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cfwtm"] Oct 03 16:37:31 crc kubenswrapper[4744]: I1003 16:37:31.984366 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cfwtm" podUID="e5b5646a-60f0-486c-8949-35b2c6965c9d" containerName="registry-server" containerID="cri-o://e9a80f926cfc432ff65fd034cb3165edb19080d65e1488767a8a90196abce6f3" gracePeriod=2 Oct 03 16:37:32 crc kubenswrapper[4744]: I1003 16:37:32.395242 4744 generic.go:334] "Generic (PLEG): container finished" podID="afc8695c-fc62-4788-9618-c0bed94eebd6" containerID="0c4a0c15f78bf7c16af9edf410bca8e20028877da4bd9f64f3f6fbea0bc65264" exitCode=0 Oct 03 16:37:32 crc kubenswrapper[4744]: I1003 16:37:32.395300 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49tm4" event={"ID":"afc8695c-fc62-4788-9618-c0bed94eebd6","Type":"ContainerDied","Data":"0c4a0c15f78bf7c16af9edf410bca8e20028877da4bd9f64f3f6fbea0bc65264"} Oct 03 16:37:32 crc kubenswrapper[4744]: I1003 16:37:32.402213 4744 generic.go:334] "Generic (PLEG): container finished" podID="e5b5646a-60f0-486c-8949-35b2c6965c9d" containerID="e9a80f926cfc432ff65fd034cb3165edb19080d65e1488767a8a90196abce6f3" exitCode=0 Oct 03 16:37:32 crc kubenswrapper[4744]: I1003 16:37:32.402270 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfwtm" event={"ID":"e5b5646a-60f0-486c-8949-35b2c6965c9d","Type":"ContainerDied","Data":"e9a80f926cfc432ff65fd034cb3165edb19080d65e1488767a8a90196abce6f3"} Oct 03 16:37:32 crc kubenswrapper[4744]: I1003 16:37:32.496366 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:32 crc kubenswrapper[4744]: I1003 16:37:32.604410 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx5mv\" (UniqueName: \"kubernetes.io/projected/e5b5646a-60f0-486c-8949-35b2c6965c9d-kube-api-access-jx5mv\") pod \"e5b5646a-60f0-486c-8949-35b2c6965c9d\" (UID: \"e5b5646a-60f0-486c-8949-35b2c6965c9d\") " Oct 03 16:37:32 crc kubenswrapper[4744]: I1003 16:37:32.604599 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5b5646a-60f0-486c-8949-35b2c6965c9d-utilities\") pod \"e5b5646a-60f0-486c-8949-35b2c6965c9d\" (UID: \"e5b5646a-60f0-486c-8949-35b2c6965c9d\") " Oct 03 16:37:32 crc kubenswrapper[4744]: I1003 16:37:32.604627 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5b5646a-60f0-486c-8949-35b2c6965c9d-catalog-content\") pod \"e5b5646a-60f0-486c-8949-35b2c6965c9d\" (UID: \"e5b5646a-60f0-486c-8949-35b2c6965c9d\") " Oct 03 16:37:32 crc kubenswrapper[4744]: I1003 16:37:32.606310 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5b5646a-60f0-486c-8949-35b2c6965c9d-utilities" (OuterVolumeSpecName: "utilities") pod "e5b5646a-60f0-486c-8949-35b2c6965c9d" (UID: "e5b5646a-60f0-486c-8949-35b2c6965c9d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:37:32 crc kubenswrapper[4744]: I1003 16:37:32.616868 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5b5646a-60f0-486c-8949-35b2c6965c9d-kube-api-access-jx5mv" (OuterVolumeSpecName: "kube-api-access-jx5mv") pod "e5b5646a-60f0-486c-8949-35b2c6965c9d" (UID: "e5b5646a-60f0-486c-8949-35b2c6965c9d"). InnerVolumeSpecName "kube-api-access-jx5mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:37:32 crc kubenswrapper[4744]: I1003 16:37:32.706717 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5b5646a-60f0-486c-8949-35b2c6965c9d-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:32 crc kubenswrapper[4744]: I1003 16:37:32.707053 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx5mv\" (UniqueName: \"kubernetes.io/projected/e5b5646a-60f0-486c-8949-35b2c6965c9d-kube-api-access-jx5mv\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:32 crc kubenswrapper[4744]: I1003 16:37:32.715336 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5b5646a-60f0-486c-8949-35b2c6965c9d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5b5646a-60f0-486c-8949-35b2c6965c9d" (UID: "e5b5646a-60f0-486c-8949-35b2c6965c9d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:37:32 crc kubenswrapper[4744]: I1003 16:37:32.808119 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5b5646a-60f0-486c-8949-35b2c6965c9d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:33 crc kubenswrapper[4744]: I1003 16:37:33.410837 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cfwtm" event={"ID":"e5b5646a-60f0-486c-8949-35b2c6965c9d","Type":"ContainerDied","Data":"20cba7b040b4bb16407dd7d3b7ba96674d5622da429097d9d1d835e431ce7c24"} Oct 03 16:37:33 crc kubenswrapper[4744]: I1003 16:37:33.410962 4744 scope.go:117] "RemoveContainer" containerID="e9a80f926cfc432ff65fd034cb3165edb19080d65e1488767a8a90196abce6f3" Oct 03 16:37:33 crc kubenswrapper[4744]: I1003 16:37:33.410866 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cfwtm" Oct 03 16:37:33 crc kubenswrapper[4744]: I1003 16:37:33.413172 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49tm4" event={"ID":"afc8695c-fc62-4788-9618-c0bed94eebd6","Type":"ContainerStarted","Data":"5b774729e2c1f10414f122a3c969613b24e8815b42a76ab588b3d365a3175bc0"} Oct 03 16:37:33 crc kubenswrapper[4744]: I1003 16:37:33.431732 4744 scope.go:117] "RemoveContainer" containerID="a857c1f14850cf7624323e8a173194275a7c5f2d7804fd5d3ef63bb425c612d8" Oct 03 16:37:33 crc kubenswrapper[4744]: I1003 16:37:33.436228 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-49tm4" podStartSLOduration=3.011861986 podStartE2EDuration="5.436214292s" podCreationTimestamp="2025-10-03 16:37:28 +0000 UTC" firstStartedPulling="2025-10-03 16:37:30.379279589 +0000 UTC m=+776.659155515" lastFinishedPulling="2025-10-03 16:37:32.803631915 +0000 UTC m=+779.083507821" observedRunningTime="2025-10-03 16:37:33.434085757 +0000 UTC m=+779.713961713" watchObservedRunningTime="2025-10-03 16:37:33.436214292 +0000 UTC m=+779.716090188" Oct 03 16:37:33 crc kubenswrapper[4744]: I1003 16:37:33.453020 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cfwtm"] Oct 03 16:37:33 crc kubenswrapper[4744]: I1003 16:37:33.457427 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cfwtm"] Oct 03 16:37:33 crc kubenswrapper[4744]: I1003 16:37:33.468878 4744 scope.go:117] "RemoveContainer" containerID="910c6a57d2590f58fb6f6726d5a5486096cb43b934c836d3f9eaed34fcd41924" Oct 03 16:37:34 crc kubenswrapper[4744]: I1003 16:37:34.669305 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:37:34 crc kubenswrapper[4744]: I1003 16:37:34.669417 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:37:34 crc kubenswrapper[4744]: I1003 16:37:34.910364 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5b5646a-60f0-486c-8949-35b2c6965c9d" path="/var/lib/kubelet/pods/e5b5646a-60f0-486c-8949-35b2c6965c9d/volumes" Oct 03 16:37:35 crc kubenswrapper[4744]: I1003 16:37:35.955949 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-pdpss"] Oct 03 16:37:35 crc kubenswrapper[4744]: E1003 16:37:35.956642 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b5646a-60f0-486c-8949-35b2c6965c9d" containerName="registry-server" Oct 03 16:37:35 crc kubenswrapper[4744]: I1003 16:37:35.956662 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b5646a-60f0-486c-8949-35b2c6965c9d" containerName="registry-server" Oct 03 16:37:35 crc kubenswrapper[4744]: E1003 16:37:35.956675 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b5646a-60f0-486c-8949-35b2c6965c9d" containerName="extract-content" Oct 03 16:37:35 crc kubenswrapper[4744]: I1003 16:37:35.956685 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b5646a-60f0-486c-8949-35b2c6965c9d" containerName="extract-content" Oct 03 16:37:35 crc kubenswrapper[4744]: E1003 16:37:35.956709 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b5646a-60f0-486c-8949-35b2c6965c9d" containerName="extract-utilities" Oct 03 16:37:35 crc kubenswrapper[4744]: I1003 16:37:35.956716 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b5646a-60f0-486c-8949-35b2c6965c9d" containerName="extract-utilities" Oct 03 16:37:35 crc kubenswrapper[4744]: I1003 16:37:35.956854 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5b5646a-60f0-486c-8949-35b2c6965c9d" containerName="registry-server" Oct 03 16:37:35 crc kubenswrapper[4744]: I1003 16:37:35.957589 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdpss" Oct 03 16:37:35 crc kubenswrapper[4744]: I1003 16:37:35.959914 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-77p8k" Oct 03 16:37:35 crc kubenswrapper[4744]: I1003 16:37:35.970446 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f"] Oct 03 16:37:35 crc kubenswrapper[4744]: I1003 16:37:35.971428 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f" Oct 03 16:37:35 crc kubenswrapper[4744]: I1003 16:37:35.973120 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 03 16:37:35 crc kubenswrapper[4744]: I1003 16:37:35.975980 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-pdpss"] Oct 03 16:37:35 crc kubenswrapper[4744]: I1003 16:37:35.999557 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-cvn5h"] Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.000325 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.009861 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f"] Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.068339 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx6cc\" (UniqueName: \"kubernetes.io/projected/4039e497-2af2-483a-8a21-14f14df6df33-kube-api-access-hx6cc\") pod \"nmstate-metrics-fdff9cb8d-pdpss\" (UID: \"4039e497-2af2-483a-8a21-14f14df6df33\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdpss" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.068526 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmk56\" (UniqueName: \"kubernetes.io/projected/63038935-53b5-4371-ab75-80e6c1522980-kube-api-access-qmk56\") pod \"nmstate-webhook-6cdbc54649-dmz4f\" (UID: \"63038935-53b5-4371-ab75-80e6c1522980\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.068564 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/63038935-53b5-4371-ab75-80e6c1522980-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-dmz4f\" (UID: \"63038935-53b5-4371-ab75-80e6c1522980\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.081748 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7"] Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.082418 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.096730 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.098244 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.099560 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-686qn" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.101318 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7"] Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.171084 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c5f8637e-bf33-4533-970d-a21b8837ce78-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-587v7\" (UID: \"c5f8637e-bf33-4533-970d-a21b8837ce78\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.171191 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx6cc\" (UniqueName: \"kubernetes.io/projected/4039e497-2af2-483a-8a21-14f14df6df33-kube-api-access-hx6cc\") pod \"nmstate-metrics-fdff9cb8d-pdpss\" (UID: \"4039e497-2af2-483a-8a21-14f14df6df33\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdpss" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.171232 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5f8637e-bf33-4533-970d-a21b8837ce78-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-587v7\" (UID: \"c5f8637e-bf33-4533-970d-a21b8837ce78\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.171306 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fcbl\" (UniqueName: \"kubernetes.io/projected/d0fe8f50-20f3-483f-8904-8f084fba1fb1-kube-api-access-8fcbl\") pod \"nmstate-handler-cvn5h\" (UID: \"d0fe8f50-20f3-483f-8904-8f084fba1fb1\") " pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.171341 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmk56\" (UniqueName: \"kubernetes.io/projected/63038935-53b5-4371-ab75-80e6c1522980-kube-api-access-qmk56\") pod \"nmstate-webhook-6cdbc54649-dmz4f\" (UID: \"63038935-53b5-4371-ab75-80e6c1522980\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.171374 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d0fe8f50-20f3-483f-8904-8f084fba1fb1-nmstate-lock\") pod \"nmstate-handler-cvn5h\" (UID: \"d0fe8f50-20f3-483f-8904-8f084fba1fb1\") " pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.171401 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d0fe8f50-20f3-483f-8904-8f084fba1fb1-ovs-socket\") pod \"nmstate-handler-cvn5h\" (UID: \"d0fe8f50-20f3-483f-8904-8f084fba1fb1\") " pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.171426 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhmx5\" (UniqueName: \"kubernetes.io/projected/c5f8637e-bf33-4533-970d-a21b8837ce78-kube-api-access-xhmx5\") pod \"nmstate-console-plugin-6b874cbd85-587v7\" (UID: \"c5f8637e-bf33-4533-970d-a21b8837ce78\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.171458 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d0fe8f50-20f3-483f-8904-8f084fba1fb1-dbus-socket\") pod \"nmstate-handler-cvn5h\" (UID: \"d0fe8f50-20f3-483f-8904-8f084fba1fb1\") " pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.171509 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/63038935-53b5-4371-ab75-80e6c1522980-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-dmz4f\" (UID: \"63038935-53b5-4371-ab75-80e6c1522980\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f" Oct 03 16:37:36 crc kubenswrapper[4744]: E1003 16:37:36.171717 4744 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 03 16:37:36 crc kubenswrapper[4744]: E1003 16:37:36.171806 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/63038935-53b5-4371-ab75-80e6c1522980-tls-key-pair podName:63038935-53b5-4371-ab75-80e6c1522980 nodeName:}" failed. No retries permitted until 2025-10-03 16:37:36.671774985 +0000 UTC m=+782.951650891 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/63038935-53b5-4371-ab75-80e6c1522980-tls-key-pair") pod "nmstate-webhook-6cdbc54649-dmz4f" (UID: "63038935-53b5-4371-ab75-80e6c1522980") : secret "openshift-nmstate-webhook" not found Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.197533 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmk56\" (UniqueName: \"kubernetes.io/projected/63038935-53b5-4371-ab75-80e6c1522980-kube-api-access-qmk56\") pod \"nmstate-webhook-6cdbc54649-dmz4f\" (UID: \"63038935-53b5-4371-ab75-80e6c1522980\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.213145 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx6cc\" (UniqueName: \"kubernetes.io/projected/4039e497-2af2-483a-8a21-14f14df6df33-kube-api-access-hx6cc\") pod \"nmstate-metrics-fdff9cb8d-pdpss\" (UID: \"4039e497-2af2-483a-8a21-14f14df6df33\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdpss" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.272098 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6fb7678fb5-f7p7s"] Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.272816 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.272843 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c5f8637e-bf33-4533-970d-a21b8837ce78-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-587v7\" (UID: \"c5f8637e-bf33-4533-970d-a21b8837ce78\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.272891 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5f8637e-bf33-4533-970d-a21b8837ce78-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-587v7\" (UID: \"c5f8637e-bf33-4533-970d-a21b8837ce78\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.272940 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fcbl\" (UniqueName: \"kubernetes.io/projected/d0fe8f50-20f3-483f-8904-8f084fba1fb1-kube-api-access-8fcbl\") pod \"nmstate-handler-cvn5h\" (UID: \"d0fe8f50-20f3-483f-8904-8f084fba1fb1\") " pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.272961 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d0fe8f50-20f3-483f-8904-8f084fba1fb1-nmstate-lock\") pod \"nmstate-handler-cvn5h\" (UID: \"d0fe8f50-20f3-483f-8904-8f084fba1fb1\") " pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.272982 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d0fe8f50-20f3-483f-8904-8f084fba1fb1-ovs-socket\") pod \"nmstate-handler-cvn5h\" (UID: \"d0fe8f50-20f3-483f-8904-8f084fba1fb1\") " pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.272999 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhmx5\" (UniqueName: \"kubernetes.io/projected/c5f8637e-bf33-4533-970d-a21b8837ce78-kube-api-access-xhmx5\") pod \"nmstate-console-plugin-6b874cbd85-587v7\" (UID: \"c5f8637e-bf33-4533-970d-a21b8837ce78\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.273016 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d0fe8f50-20f3-483f-8904-8f084fba1fb1-dbus-socket\") pod \"nmstate-handler-cvn5h\" (UID: \"d0fe8f50-20f3-483f-8904-8f084fba1fb1\") " pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.273171 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d0fe8f50-20f3-483f-8904-8f084fba1fb1-ovs-socket\") pod \"nmstate-handler-cvn5h\" (UID: \"d0fe8f50-20f3-483f-8904-8f084fba1fb1\") " pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.273231 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d0fe8f50-20f3-483f-8904-8f084fba1fb1-nmstate-lock\") pod \"nmstate-handler-cvn5h\" (UID: \"d0fe8f50-20f3-483f-8904-8f084fba1fb1\") " pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.273326 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d0fe8f50-20f3-483f-8904-8f084fba1fb1-dbus-socket\") pod \"nmstate-handler-cvn5h\" (UID: \"d0fe8f50-20f3-483f-8904-8f084fba1fb1\") " pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.273984 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c5f8637e-bf33-4533-970d-a21b8837ce78-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-587v7\" (UID: \"c5f8637e-bf33-4533-970d-a21b8837ce78\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.277447 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdpss" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.289206 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhmx5\" (UniqueName: \"kubernetes.io/projected/c5f8637e-bf33-4533-970d-a21b8837ce78-kube-api-access-xhmx5\") pod \"nmstate-console-plugin-6b874cbd85-587v7\" (UID: \"c5f8637e-bf33-4533-970d-a21b8837ce78\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.295620 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5f8637e-bf33-4533-970d-a21b8837ce78-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-587v7\" (UID: \"c5f8637e-bf33-4533-970d-a21b8837ce78\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.296799 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6fb7678fb5-f7p7s"] Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.317509 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fcbl\" (UniqueName: \"kubernetes.io/projected/d0fe8f50-20f3-483f-8904-8f084fba1fb1-kube-api-access-8fcbl\") pod \"nmstate-handler-cvn5h\" (UID: \"d0fe8f50-20f3-483f-8904-8f084fba1fb1\") " pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.321178 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.374831 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1542b010-070f-43b6-92b9-107ea16631a6-oauth-serving-cert\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.374883 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1542b010-070f-43b6-92b9-107ea16631a6-console-oauth-config\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.374920 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1542b010-070f-43b6-92b9-107ea16631a6-trusted-ca-bundle\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.374945 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hj8s\" (UniqueName: \"kubernetes.io/projected/1542b010-070f-43b6-92b9-107ea16631a6-kube-api-access-7hj8s\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.375013 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1542b010-070f-43b6-92b9-107ea16631a6-console-serving-cert\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.375045 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1542b010-070f-43b6-92b9-107ea16631a6-console-config\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.375087 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1542b010-070f-43b6-92b9-107ea16631a6-service-ca\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.400304 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.452854 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-cvn5h" event={"ID":"d0fe8f50-20f3-483f-8904-8f084fba1fb1","Type":"ContainerStarted","Data":"7663caa302c8ab1bb27c60a9b57378ca267080a0c374c5671ede3364f742355a"} Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.476673 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1542b010-070f-43b6-92b9-107ea16631a6-console-serving-cert\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.476722 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1542b010-070f-43b6-92b9-107ea16631a6-console-config\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.476757 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1542b010-070f-43b6-92b9-107ea16631a6-service-ca\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.476822 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1542b010-070f-43b6-92b9-107ea16631a6-oauth-serving-cert\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.476841 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1542b010-070f-43b6-92b9-107ea16631a6-console-oauth-config\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.476867 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1542b010-070f-43b6-92b9-107ea16631a6-trusted-ca-bundle\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.476886 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hj8s\" (UniqueName: \"kubernetes.io/projected/1542b010-070f-43b6-92b9-107ea16631a6-kube-api-access-7hj8s\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.478594 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1542b010-070f-43b6-92b9-107ea16631a6-oauth-serving-cert\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.478612 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1542b010-070f-43b6-92b9-107ea16631a6-service-ca\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.478655 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1542b010-070f-43b6-92b9-107ea16631a6-trusted-ca-bundle\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.478682 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1542b010-070f-43b6-92b9-107ea16631a6-console-config\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.482907 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1542b010-070f-43b6-92b9-107ea16631a6-console-oauth-config\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.486289 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1542b010-070f-43b6-92b9-107ea16631a6-console-serving-cert\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.492110 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hj8s\" (UniqueName: \"kubernetes.io/projected/1542b010-070f-43b6-92b9-107ea16631a6-kube-api-access-7hj8s\") pod \"console-6fb7678fb5-f7p7s\" (UID: \"1542b010-070f-43b6-92b9-107ea16631a6\") " pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.653866 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.680201 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/63038935-53b5-4371-ab75-80e6c1522980-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-dmz4f\" (UID: \"63038935-53b5-4371-ab75-80e6c1522980\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.685946 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/63038935-53b5-4371-ab75-80e6c1522980-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-dmz4f\" (UID: \"63038935-53b5-4371-ab75-80e6c1522980\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f" Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.723142 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-pdpss"] Oct 03 16:37:36 crc kubenswrapper[4744]: W1003 16:37:36.734956 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4039e497_2af2_483a_8a21_14f14df6df33.slice/crio-17a8b87c3c66e7d2d0a25ff8f0039e92acf12fa8966acd5e5af17370ab8cf718 WatchSource:0}: Error finding container 17a8b87c3c66e7d2d0a25ff8f0039e92acf12fa8966acd5e5af17370ab8cf718: Status 404 returned error can't find the container with id 17a8b87c3c66e7d2d0a25ff8f0039e92acf12fa8966acd5e5af17370ab8cf718 Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.812246 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7"] Oct 03 16:37:36 crc kubenswrapper[4744]: W1003 16:37:36.816378 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5f8637e_bf33_4533_970d_a21b8837ce78.slice/crio-41bc47f34f92a5949bd294f8a277abe9043d2fce63beafb1bbfd28ee8d6d40b1 WatchSource:0}: Error finding container 41bc47f34f92a5949bd294f8a277abe9043d2fce63beafb1bbfd28ee8d6d40b1: Status 404 returned error can't find the container with id 41bc47f34f92a5949bd294f8a277abe9043d2fce63beafb1bbfd28ee8d6d40b1 Oct 03 16:37:36 crc kubenswrapper[4744]: I1003 16:37:36.893615 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f" Oct 03 16:37:37 crc kubenswrapper[4744]: I1003 16:37:37.115609 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6fb7678fb5-f7p7s"] Oct 03 16:37:37 crc kubenswrapper[4744]: W1003 16:37:37.121643 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1542b010_070f_43b6_92b9_107ea16631a6.slice/crio-c800b3d33cb87ccdf3330abdde107060f0df40a8aa225a1b4f4bde4947adb692 WatchSource:0}: Error finding container c800b3d33cb87ccdf3330abdde107060f0df40a8aa225a1b4f4bde4947adb692: Status 404 returned error can't find the container with id c800b3d33cb87ccdf3330abdde107060f0df40a8aa225a1b4f4bde4947adb692 Oct 03 16:37:37 crc kubenswrapper[4744]: I1003 16:37:37.299727 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f"] Oct 03 16:37:37 crc kubenswrapper[4744]: W1003 16:37:37.305572 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63038935_53b5_4371_ab75_80e6c1522980.slice/crio-5324cdafc0c7e21418a950d32d219c79e2ef9e5662d0f0422cd63396c4c10abb WatchSource:0}: Error finding container 5324cdafc0c7e21418a950d32d219c79e2ef9e5662d0f0422cd63396c4c10abb: Status 404 returned error can't find the container with id 5324cdafc0c7e21418a950d32d219c79e2ef9e5662d0f0422cd63396c4c10abb Oct 03 16:37:37 crc kubenswrapper[4744]: I1003 16:37:37.459826 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdpss" event={"ID":"4039e497-2af2-483a-8a21-14f14df6df33","Type":"ContainerStarted","Data":"17a8b87c3c66e7d2d0a25ff8f0039e92acf12fa8966acd5e5af17370ab8cf718"} Oct 03 16:37:37 crc kubenswrapper[4744]: I1003 16:37:37.461819 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f" event={"ID":"63038935-53b5-4371-ab75-80e6c1522980","Type":"ContainerStarted","Data":"5324cdafc0c7e21418a950d32d219c79e2ef9e5662d0f0422cd63396c4c10abb"} Oct 03 16:37:37 crc kubenswrapper[4744]: I1003 16:37:37.463424 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7" event={"ID":"c5f8637e-bf33-4533-970d-a21b8837ce78","Type":"ContainerStarted","Data":"41bc47f34f92a5949bd294f8a277abe9043d2fce63beafb1bbfd28ee8d6d40b1"} Oct 03 16:37:37 crc kubenswrapper[4744]: I1003 16:37:37.465859 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6fb7678fb5-f7p7s" event={"ID":"1542b010-070f-43b6-92b9-107ea16631a6","Type":"ContainerStarted","Data":"f2709fdddd190c5b91732d0ce084b46adb690432289871260d4a7f54503f7af2"} Oct 03 16:37:37 crc kubenswrapper[4744]: I1003 16:37:37.465922 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6fb7678fb5-f7p7s" event={"ID":"1542b010-070f-43b6-92b9-107ea16631a6","Type":"ContainerStarted","Data":"c800b3d33cb87ccdf3330abdde107060f0df40a8aa225a1b4f4bde4947adb692"} Oct 03 16:37:37 crc kubenswrapper[4744]: I1003 16:37:37.487099 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6fb7678fb5-f7p7s" podStartSLOduration=1.487053349 podStartE2EDuration="1.487053349s" podCreationTimestamp="2025-10-03 16:37:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:37:37.482798319 +0000 UTC m=+783.762674225" watchObservedRunningTime="2025-10-03 16:37:37.487053349 +0000 UTC m=+783.766929245" Oct 03 16:37:39 crc kubenswrapper[4744]: I1003 16:37:39.226756 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:39 crc kubenswrapper[4744]: I1003 16:37:39.226799 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:39 crc kubenswrapper[4744]: I1003 16:37:39.287825 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:39 crc kubenswrapper[4744]: I1003 16:37:39.482584 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7" event={"ID":"c5f8637e-bf33-4533-970d-a21b8837ce78","Type":"ContainerStarted","Data":"8e3fcdb94a6682d227c6a28c23e63982c23a30ec505c56bae894ad7906497461"} Oct 03 16:37:39 crc kubenswrapper[4744]: I1003 16:37:39.484101 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdpss" event={"ID":"4039e497-2af2-483a-8a21-14f14df6df33","Type":"ContainerStarted","Data":"ff827aca23f46fe1f0cf5db73b7d2f63888ce0377ef16070dae3a7a7317d3204"} Oct 03 16:37:39 crc kubenswrapper[4744]: I1003 16:37:39.501444 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-587v7" podStartSLOduration=1.12385982 podStartE2EDuration="3.501427022s" podCreationTimestamp="2025-10-03 16:37:36 +0000 UTC" firstStartedPulling="2025-10-03 16:37:36.818889575 +0000 UTC m=+783.098765471" lastFinishedPulling="2025-10-03 16:37:39.196456767 +0000 UTC m=+785.476332673" observedRunningTime="2025-10-03 16:37:39.498087146 +0000 UTC m=+785.777963042" watchObservedRunningTime="2025-10-03 16:37:39.501427022 +0000 UTC m=+785.781302918" Oct 03 16:37:39 crc kubenswrapper[4744]: I1003 16:37:39.535328 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:39 crc kubenswrapper[4744]: I1003 16:37:39.585961 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-49tm4"] Oct 03 16:37:40 crc kubenswrapper[4744]: I1003 16:37:40.490961 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f" event={"ID":"63038935-53b5-4371-ab75-80e6c1522980","Type":"ContainerStarted","Data":"a0581c36d3d6f13bc272fe381f982d09e2589f0139390cabe4efc36b1f9f2ab0"} Oct 03 16:37:40 crc kubenswrapper[4744]: I1003 16:37:40.491990 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f" Oct 03 16:37:40 crc kubenswrapper[4744]: I1003 16:37:40.492410 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-cvn5h" event={"ID":"d0fe8f50-20f3-483f-8904-8f084fba1fb1","Type":"ContainerStarted","Data":"0c1e061ca4c379c370e19a85d48378fa90010e38c50adab2793dcbd6ae803478"} Oct 03 16:37:40 crc kubenswrapper[4744]: I1003 16:37:40.510830 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f" podStartSLOduration=3.618382203 podStartE2EDuration="5.510793676s" podCreationTimestamp="2025-10-03 16:37:35 +0000 UTC" firstStartedPulling="2025-10-03 16:37:37.307321948 +0000 UTC m=+783.587197844" lastFinishedPulling="2025-10-03 16:37:39.199733421 +0000 UTC m=+785.479609317" observedRunningTime="2025-10-03 16:37:40.50784681 +0000 UTC m=+786.787722706" watchObservedRunningTime="2025-10-03 16:37:40.510793676 +0000 UTC m=+786.790669572" Oct 03 16:37:40 crc kubenswrapper[4744]: I1003 16:37:40.533652 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-cvn5h" podStartSLOduration=2.688035286 podStartE2EDuration="5.533626014s" podCreationTimestamp="2025-10-03 16:37:35 +0000 UTC" firstStartedPulling="2025-10-03 16:37:36.351456044 +0000 UTC m=+782.631331940" lastFinishedPulling="2025-10-03 16:37:39.197046772 +0000 UTC m=+785.476922668" observedRunningTime="2025-10-03 16:37:40.532977447 +0000 UTC m=+786.812853343" watchObservedRunningTime="2025-10-03 16:37:40.533626014 +0000 UTC m=+786.813501930" Oct 03 16:37:41 crc kubenswrapper[4744]: I1003 16:37:41.327447 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:41 crc kubenswrapper[4744]: I1003 16:37:41.497465 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-49tm4" podUID="afc8695c-fc62-4788-9618-c0bed94eebd6" containerName="registry-server" containerID="cri-o://5b774729e2c1f10414f122a3c969613b24e8815b42a76ab588b3d365a3175bc0" gracePeriod=2 Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.109869 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.266970 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8l7hz\" (UniqueName: \"kubernetes.io/projected/afc8695c-fc62-4788-9618-c0bed94eebd6-kube-api-access-8l7hz\") pod \"afc8695c-fc62-4788-9618-c0bed94eebd6\" (UID: \"afc8695c-fc62-4788-9618-c0bed94eebd6\") " Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.267064 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc8695c-fc62-4788-9618-c0bed94eebd6-utilities\") pod \"afc8695c-fc62-4788-9618-c0bed94eebd6\" (UID: \"afc8695c-fc62-4788-9618-c0bed94eebd6\") " Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.267236 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc8695c-fc62-4788-9618-c0bed94eebd6-catalog-content\") pod \"afc8695c-fc62-4788-9618-c0bed94eebd6\" (UID: \"afc8695c-fc62-4788-9618-c0bed94eebd6\") " Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.268008 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afc8695c-fc62-4788-9618-c0bed94eebd6-utilities" (OuterVolumeSpecName: "utilities") pod "afc8695c-fc62-4788-9618-c0bed94eebd6" (UID: "afc8695c-fc62-4788-9618-c0bed94eebd6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.272125 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc8695c-fc62-4788-9618-c0bed94eebd6-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.283702 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afc8695c-fc62-4788-9618-c0bed94eebd6-kube-api-access-8l7hz" (OuterVolumeSpecName: "kube-api-access-8l7hz") pod "afc8695c-fc62-4788-9618-c0bed94eebd6" (UID: "afc8695c-fc62-4788-9618-c0bed94eebd6"). InnerVolumeSpecName "kube-api-access-8l7hz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.316174 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afc8695c-fc62-4788-9618-c0bed94eebd6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "afc8695c-fc62-4788-9618-c0bed94eebd6" (UID: "afc8695c-fc62-4788-9618-c0bed94eebd6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.372923 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc8695c-fc62-4788-9618-c0bed94eebd6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.372960 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8l7hz\" (UniqueName: \"kubernetes.io/projected/afc8695c-fc62-4788-9618-c0bed94eebd6-kube-api-access-8l7hz\") on node \"crc\" DevicePath \"\"" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.503163 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdpss" event={"ID":"4039e497-2af2-483a-8a21-14f14df6df33","Type":"ContainerStarted","Data":"e3b61a042b867f720b511bdcdd454c4d151077a35ae61a1075e1391c5c8acad5"} Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.505403 4744 generic.go:334] "Generic (PLEG): container finished" podID="afc8695c-fc62-4788-9618-c0bed94eebd6" containerID="5b774729e2c1f10414f122a3c969613b24e8815b42a76ab588b3d365a3175bc0" exitCode=0 Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.505511 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49tm4" event={"ID":"afc8695c-fc62-4788-9618-c0bed94eebd6","Type":"ContainerDied","Data":"5b774729e2c1f10414f122a3c969613b24e8815b42a76ab588b3d365a3175bc0"} Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.505521 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49tm4" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.505548 4744 scope.go:117] "RemoveContainer" containerID="5b774729e2c1f10414f122a3c969613b24e8815b42a76ab588b3d365a3175bc0" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.505534 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49tm4" event={"ID":"afc8695c-fc62-4788-9618-c0bed94eebd6","Type":"ContainerDied","Data":"056b2eb320c5e152472bdda8b46b48f0a929303011708b50b20929dedbc8dd30"} Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.527929 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdpss" podStartSLOduration=2.459721285 podStartE2EDuration="7.5278991s" podCreationTimestamp="2025-10-03 16:37:35 +0000 UTC" firstStartedPulling="2025-10-03 16:37:36.73791976 +0000 UTC m=+783.017795666" lastFinishedPulling="2025-10-03 16:37:41.806097565 +0000 UTC m=+788.085973481" observedRunningTime="2025-10-03 16:37:42.518996011 +0000 UTC m=+788.798871927" watchObservedRunningTime="2025-10-03 16:37:42.5278991 +0000 UTC m=+788.807775036" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.541614 4744 scope.go:117] "RemoveContainer" containerID="0c4a0c15f78bf7c16af9edf410bca8e20028877da4bd9f64f3f6fbea0bc65264" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.548430 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-49tm4"] Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.560274 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-49tm4"] Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.564034 4744 scope.go:117] "RemoveContainer" containerID="937eb220f563d0f6f9dfa1e1b0b7324a4c48241f3bcabc6fa993e330a6d7f78d" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.581993 4744 scope.go:117] "RemoveContainer" containerID="5b774729e2c1f10414f122a3c969613b24e8815b42a76ab588b3d365a3175bc0" Oct 03 16:37:42 crc kubenswrapper[4744]: E1003 16:37:42.582791 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b774729e2c1f10414f122a3c969613b24e8815b42a76ab588b3d365a3175bc0\": container with ID starting with 5b774729e2c1f10414f122a3c969613b24e8815b42a76ab588b3d365a3175bc0 not found: ID does not exist" containerID="5b774729e2c1f10414f122a3c969613b24e8815b42a76ab588b3d365a3175bc0" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.582835 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b774729e2c1f10414f122a3c969613b24e8815b42a76ab588b3d365a3175bc0"} err="failed to get container status \"5b774729e2c1f10414f122a3c969613b24e8815b42a76ab588b3d365a3175bc0\": rpc error: code = NotFound desc = could not find container \"5b774729e2c1f10414f122a3c969613b24e8815b42a76ab588b3d365a3175bc0\": container with ID starting with 5b774729e2c1f10414f122a3c969613b24e8815b42a76ab588b3d365a3175bc0 not found: ID does not exist" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.582861 4744 scope.go:117] "RemoveContainer" containerID="0c4a0c15f78bf7c16af9edf410bca8e20028877da4bd9f64f3f6fbea0bc65264" Oct 03 16:37:42 crc kubenswrapper[4744]: E1003 16:37:42.583157 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c4a0c15f78bf7c16af9edf410bca8e20028877da4bd9f64f3f6fbea0bc65264\": container with ID starting with 0c4a0c15f78bf7c16af9edf410bca8e20028877da4bd9f64f3f6fbea0bc65264 not found: ID does not exist" containerID="0c4a0c15f78bf7c16af9edf410bca8e20028877da4bd9f64f3f6fbea0bc65264" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.583189 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c4a0c15f78bf7c16af9edf410bca8e20028877da4bd9f64f3f6fbea0bc65264"} err="failed to get container status \"0c4a0c15f78bf7c16af9edf410bca8e20028877da4bd9f64f3f6fbea0bc65264\": rpc error: code = NotFound desc = could not find container \"0c4a0c15f78bf7c16af9edf410bca8e20028877da4bd9f64f3f6fbea0bc65264\": container with ID starting with 0c4a0c15f78bf7c16af9edf410bca8e20028877da4bd9f64f3f6fbea0bc65264 not found: ID does not exist" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.583204 4744 scope.go:117] "RemoveContainer" containerID="937eb220f563d0f6f9dfa1e1b0b7324a4c48241f3bcabc6fa993e330a6d7f78d" Oct 03 16:37:42 crc kubenswrapper[4744]: E1003 16:37:42.583413 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"937eb220f563d0f6f9dfa1e1b0b7324a4c48241f3bcabc6fa993e330a6d7f78d\": container with ID starting with 937eb220f563d0f6f9dfa1e1b0b7324a4c48241f3bcabc6fa993e330a6d7f78d not found: ID does not exist" containerID="937eb220f563d0f6f9dfa1e1b0b7324a4c48241f3bcabc6fa993e330a6d7f78d" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.583445 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"937eb220f563d0f6f9dfa1e1b0b7324a4c48241f3bcabc6fa993e330a6d7f78d"} err="failed to get container status \"937eb220f563d0f6f9dfa1e1b0b7324a4c48241f3bcabc6fa993e330a6d7f78d\": rpc error: code = NotFound desc = could not find container \"937eb220f563d0f6f9dfa1e1b0b7324a4c48241f3bcabc6fa993e330a6d7f78d\": container with ID starting with 937eb220f563d0f6f9dfa1e1b0b7324a4c48241f3bcabc6fa993e330a6d7f78d not found: ID does not exist" Oct 03 16:37:42 crc kubenswrapper[4744]: I1003 16:37:42.906553 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afc8695c-fc62-4788-9618-c0bed94eebd6" path="/var/lib/kubelet/pods/afc8695c-fc62-4788-9618-c0bed94eebd6/volumes" Oct 03 16:37:46 crc kubenswrapper[4744]: I1003 16:37:46.372955 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-cvn5h" Oct 03 16:37:46 crc kubenswrapper[4744]: I1003 16:37:46.654647 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:46 crc kubenswrapper[4744]: I1003 16:37:46.654739 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:46 crc kubenswrapper[4744]: I1003 16:37:46.663484 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:47 crc kubenswrapper[4744]: I1003 16:37:47.559219 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6fb7678fb5-f7p7s" Oct 03 16:37:47 crc kubenswrapper[4744]: I1003 16:37:47.660326 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-v5snt"] Oct 03 16:37:55 crc kubenswrapper[4744]: I1003 16:37:55.850799 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r58bx"] Oct 03 16:37:55 crc kubenswrapper[4744]: E1003 16:37:55.851761 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc8695c-fc62-4788-9618-c0bed94eebd6" containerName="registry-server" Oct 03 16:37:55 crc kubenswrapper[4744]: I1003 16:37:55.851779 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc8695c-fc62-4788-9618-c0bed94eebd6" containerName="registry-server" Oct 03 16:37:55 crc kubenswrapper[4744]: E1003 16:37:55.851800 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc8695c-fc62-4788-9618-c0bed94eebd6" containerName="extract-utilities" Oct 03 16:37:55 crc kubenswrapper[4744]: I1003 16:37:55.851808 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc8695c-fc62-4788-9618-c0bed94eebd6" containerName="extract-utilities" Oct 03 16:37:55 crc kubenswrapper[4744]: E1003 16:37:55.851829 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc8695c-fc62-4788-9618-c0bed94eebd6" containerName="extract-content" Oct 03 16:37:55 crc kubenswrapper[4744]: I1003 16:37:55.851838 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc8695c-fc62-4788-9618-c0bed94eebd6" containerName="extract-content" Oct 03 16:37:55 crc kubenswrapper[4744]: I1003 16:37:55.851978 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc8695c-fc62-4788-9618-c0bed94eebd6" containerName="registry-server" Oct 03 16:37:55 crc kubenswrapper[4744]: I1003 16:37:55.853003 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:37:55 crc kubenswrapper[4744]: I1003 16:37:55.876088 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r58bx"] Oct 03 16:37:55 crc kubenswrapper[4744]: I1003 16:37:55.986696 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c0ed22e-7dce-40f3-8221-d715419dbcfa-utilities\") pod \"redhat-marketplace-r58bx\" (UID: \"4c0ed22e-7dce-40f3-8221-d715419dbcfa\") " pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:37:55 crc kubenswrapper[4744]: I1003 16:37:55.986757 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr9n4\" (UniqueName: \"kubernetes.io/projected/4c0ed22e-7dce-40f3-8221-d715419dbcfa-kube-api-access-fr9n4\") pod \"redhat-marketplace-r58bx\" (UID: \"4c0ed22e-7dce-40f3-8221-d715419dbcfa\") " pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:37:55 crc kubenswrapper[4744]: I1003 16:37:55.986824 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c0ed22e-7dce-40f3-8221-d715419dbcfa-catalog-content\") pod \"redhat-marketplace-r58bx\" (UID: \"4c0ed22e-7dce-40f3-8221-d715419dbcfa\") " pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:37:56 crc kubenswrapper[4744]: I1003 16:37:56.088058 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c0ed22e-7dce-40f3-8221-d715419dbcfa-utilities\") pod \"redhat-marketplace-r58bx\" (UID: \"4c0ed22e-7dce-40f3-8221-d715419dbcfa\") " pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:37:56 crc kubenswrapper[4744]: I1003 16:37:56.088113 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr9n4\" (UniqueName: \"kubernetes.io/projected/4c0ed22e-7dce-40f3-8221-d715419dbcfa-kube-api-access-fr9n4\") pod \"redhat-marketplace-r58bx\" (UID: \"4c0ed22e-7dce-40f3-8221-d715419dbcfa\") " pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:37:56 crc kubenswrapper[4744]: I1003 16:37:56.088159 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c0ed22e-7dce-40f3-8221-d715419dbcfa-catalog-content\") pod \"redhat-marketplace-r58bx\" (UID: \"4c0ed22e-7dce-40f3-8221-d715419dbcfa\") " pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:37:56 crc kubenswrapper[4744]: I1003 16:37:56.088746 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c0ed22e-7dce-40f3-8221-d715419dbcfa-utilities\") pod \"redhat-marketplace-r58bx\" (UID: \"4c0ed22e-7dce-40f3-8221-d715419dbcfa\") " pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:37:56 crc kubenswrapper[4744]: I1003 16:37:56.088753 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c0ed22e-7dce-40f3-8221-d715419dbcfa-catalog-content\") pod \"redhat-marketplace-r58bx\" (UID: \"4c0ed22e-7dce-40f3-8221-d715419dbcfa\") " pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:37:56 crc kubenswrapper[4744]: I1003 16:37:56.111655 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr9n4\" (UniqueName: \"kubernetes.io/projected/4c0ed22e-7dce-40f3-8221-d715419dbcfa-kube-api-access-fr9n4\") pod \"redhat-marketplace-r58bx\" (UID: \"4c0ed22e-7dce-40f3-8221-d715419dbcfa\") " pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:37:56 crc kubenswrapper[4744]: I1003 16:37:56.182553 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:37:56 crc kubenswrapper[4744]: I1003 16:37:56.607821 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r58bx"] Oct 03 16:37:56 crc kubenswrapper[4744]: W1003 16:37:56.621095 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c0ed22e_7dce_40f3_8221_d715419dbcfa.slice/crio-c7c00cddc025d9cbbf6d1bd85fb284d1e4894bddd2419dbdc446531c91aea7c4 WatchSource:0}: Error finding container c7c00cddc025d9cbbf6d1bd85fb284d1e4894bddd2419dbdc446531c91aea7c4: Status 404 returned error can't find the container with id c7c00cddc025d9cbbf6d1bd85fb284d1e4894bddd2419dbdc446531c91aea7c4 Oct 03 16:37:56 crc kubenswrapper[4744]: I1003 16:37:56.901120 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-dmz4f" Oct 03 16:37:57 crc kubenswrapper[4744]: I1003 16:37:57.637111 4744 generic.go:334] "Generic (PLEG): container finished" podID="4c0ed22e-7dce-40f3-8221-d715419dbcfa" containerID="d769a017abc9b678473d925f95f5916af38403609b44bfc364c8cd6d9eb8d5d7" exitCode=0 Oct 03 16:37:57 crc kubenswrapper[4744]: I1003 16:37:57.637320 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r58bx" event={"ID":"4c0ed22e-7dce-40f3-8221-d715419dbcfa","Type":"ContainerDied","Data":"d769a017abc9b678473d925f95f5916af38403609b44bfc364c8cd6d9eb8d5d7"} Oct 03 16:37:57 crc kubenswrapper[4744]: I1003 16:37:57.637906 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r58bx" event={"ID":"4c0ed22e-7dce-40f3-8221-d715419dbcfa","Type":"ContainerStarted","Data":"c7c00cddc025d9cbbf6d1bd85fb284d1e4894bddd2419dbdc446531c91aea7c4"} Oct 03 16:37:58 crc kubenswrapper[4744]: I1003 16:37:58.632219 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2k6kz"] Oct 03 16:37:58 crc kubenswrapper[4744]: I1003 16:37:58.634006 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:37:58 crc kubenswrapper[4744]: I1003 16:37:58.643685 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2k6kz"] Oct 03 16:37:58 crc kubenswrapper[4744]: I1003 16:37:58.650265 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r58bx" event={"ID":"4c0ed22e-7dce-40f3-8221-d715419dbcfa","Type":"ContainerStarted","Data":"dc5de813cefa3ea24877c079e128145fb002dca798cdcca34355d7a7278e056f"} Oct 03 16:37:58 crc kubenswrapper[4744]: I1003 16:37:58.826904 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-catalog-content\") pod \"certified-operators-2k6kz\" (UID: \"c840ac6c-28ba-4c36-aa15-30dbab14a0d7\") " pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:37:58 crc kubenswrapper[4744]: I1003 16:37:58.827037 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckhcb\" (UniqueName: \"kubernetes.io/projected/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-kube-api-access-ckhcb\") pod \"certified-operators-2k6kz\" (UID: \"c840ac6c-28ba-4c36-aa15-30dbab14a0d7\") " pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:37:58 crc kubenswrapper[4744]: I1003 16:37:58.827463 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-utilities\") pod \"certified-operators-2k6kz\" (UID: \"c840ac6c-28ba-4c36-aa15-30dbab14a0d7\") " pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:37:58 crc kubenswrapper[4744]: I1003 16:37:58.928620 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-catalog-content\") pod \"certified-operators-2k6kz\" (UID: \"c840ac6c-28ba-4c36-aa15-30dbab14a0d7\") " pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:37:58 crc kubenswrapper[4744]: I1003 16:37:58.928703 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckhcb\" (UniqueName: \"kubernetes.io/projected/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-kube-api-access-ckhcb\") pod \"certified-operators-2k6kz\" (UID: \"c840ac6c-28ba-4c36-aa15-30dbab14a0d7\") " pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:37:58 crc kubenswrapper[4744]: I1003 16:37:58.928774 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-utilities\") pod \"certified-operators-2k6kz\" (UID: \"c840ac6c-28ba-4c36-aa15-30dbab14a0d7\") " pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:37:58 crc kubenswrapper[4744]: I1003 16:37:58.929214 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-catalog-content\") pod \"certified-operators-2k6kz\" (UID: \"c840ac6c-28ba-4c36-aa15-30dbab14a0d7\") " pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:37:58 crc kubenswrapper[4744]: I1003 16:37:58.929243 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-utilities\") pod \"certified-operators-2k6kz\" (UID: \"c840ac6c-28ba-4c36-aa15-30dbab14a0d7\") " pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:37:58 crc kubenswrapper[4744]: I1003 16:37:58.958167 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckhcb\" (UniqueName: \"kubernetes.io/projected/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-kube-api-access-ckhcb\") pod \"certified-operators-2k6kz\" (UID: \"c840ac6c-28ba-4c36-aa15-30dbab14a0d7\") " pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:37:59 crc kubenswrapper[4744]: I1003 16:37:59.250125 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:37:59 crc kubenswrapper[4744]: I1003 16:37:59.659808 4744 generic.go:334] "Generic (PLEG): container finished" podID="4c0ed22e-7dce-40f3-8221-d715419dbcfa" containerID="dc5de813cefa3ea24877c079e128145fb002dca798cdcca34355d7a7278e056f" exitCode=0 Oct 03 16:37:59 crc kubenswrapper[4744]: I1003 16:37:59.659883 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r58bx" event={"ID":"4c0ed22e-7dce-40f3-8221-d715419dbcfa","Type":"ContainerDied","Data":"dc5de813cefa3ea24877c079e128145fb002dca798cdcca34355d7a7278e056f"} Oct 03 16:37:59 crc kubenswrapper[4744]: I1003 16:37:59.705171 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2k6kz"] Oct 03 16:37:59 crc kubenswrapper[4744]: W1003 16:37:59.715852 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc840ac6c_28ba_4c36_aa15_30dbab14a0d7.slice/crio-0827bc2bc1aa41df77d0e130b98d98afa69c47eaf47a3c11470df5cc2d00596f WatchSource:0}: Error finding container 0827bc2bc1aa41df77d0e130b98d98afa69c47eaf47a3c11470df5cc2d00596f: Status 404 returned error can't find the container with id 0827bc2bc1aa41df77d0e130b98d98afa69c47eaf47a3c11470df5cc2d00596f Oct 03 16:38:00 crc kubenswrapper[4744]: I1003 16:38:00.668459 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r58bx" event={"ID":"4c0ed22e-7dce-40f3-8221-d715419dbcfa","Type":"ContainerStarted","Data":"c0de36770362e004f318febc7ba3a9330fbfc0577f062503427a0bab88b942bf"} Oct 03 16:38:00 crc kubenswrapper[4744]: I1003 16:38:00.670759 4744 generic.go:334] "Generic (PLEG): container finished" podID="c840ac6c-28ba-4c36-aa15-30dbab14a0d7" containerID="32536b28176c56f4dce2657605ef79a68be4e83f283e551eaead23d8f7ad0cb5" exitCode=0 Oct 03 16:38:00 crc kubenswrapper[4744]: I1003 16:38:00.670801 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2k6kz" event={"ID":"c840ac6c-28ba-4c36-aa15-30dbab14a0d7","Type":"ContainerDied","Data":"32536b28176c56f4dce2657605ef79a68be4e83f283e551eaead23d8f7ad0cb5"} Oct 03 16:38:00 crc kubenswrapper[4744]: I1003 16:38:00.670826 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2k6kz" event={"ID":"c840ac6c-28ba-4c36-aa15-30dbab14a0d7","Type":"ContainerStarted","Data":"0827bc2bc1aa41df77d0e130b98d98afa69c47eaf47a3c11470df5cc2d00596f"} Oct 03 16:38:00 crc kubenswrapper[4744]: I1003 16:38:00.690923 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r58bx" podStartSLOduration=3.240767573 podStartE2EDuration="5.690901553s" podCreationTimestamp="2025-10-03 16:37:55 +0000 UTC" firstStartedPulling="2025-10-03 16:37:57.640701614 +0000 UTC m=+803.920577550" lastFinishedPulling="2025-10-03 16:38:00.090835624 +0000 UTC m=+806.370711530" observedRunningTime="2025-10-03 16:38:00.687646239 +0000 UTC m=+806.967522145" watchObservedRunningTime="2025-10-03 16:38:00.690901553 +0000 UTC m=+806.970777449" Oct 03 16:38:01 crc kubenswrapper[4744]: I1003 16:38:01.678913 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2k6kz" event={"ID":"c840ac6c-28ba-4c36-aa15-30dbab14a0d7","Type":"ContainerStarted","Data":"1e4e882ebe160e3c898d682ac297051575b4f051d09371fee97dc07d542fb609"} Oct 03 16:38:02 crc kubenswrapper[4744]: I1003 16:38:02.688907 4744 generic.go:334] "Generic (PLEG): container finished" podID="c840ac6c-28ba-4c36-aa15-30dbab14a0d7" containerID="1e4e882ebe160e3c898d682ac297051575b4f051d09371fee97dc07d542fb609" exitCode=0 Oct 03 16:38:02 crc kubenswrapper[4744]: I1003 16:38:02.688980 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2k6kz" event={"ID":"c840ac6c-28ba-4c36-aa15-30dbab14a0d7","Type":"ContainerDied","Data":"1e4e882ebe160e3c898d682ac297051575b4f051d09371fee97dc07d542fb609"} Oct 03 16:38:03 crc kubenswrapper[4744]: I1003 16:38:03.699898 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2k6kz" event={"ID":"c840ac6c-28ba-4c36-aa15-30dbab14a0d7","Type":"ContainerStarted","Data":"1d225d14fd98ae0698cf45e6ab542eb826cc5ee082785de0a049ce816e5d59f0"} Oct 03 16:38:03 crc kubenswrapper[4744]: I1003 16:38:03.725548 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2k6kz" podStartSLOduration=3.3286295519999998 podStartE2EDuration="5.72551452s" podCreationTimestamp="2025-10-03 16:37:58 +0000 UTC" firstStartedPulling="2025-10-03 16:38:00.671829092 +0000 UTC m=+806.951704988" lastFinishedPulling="2025-10-03 16:38:03.06871405 +0000 UTC m=+809.348589956" observedRunningTime="2025-10-03 16:38:03.722992415 +0000 UTC m=+810.002868371" watchObservedRunningTime="2025-10-03 16:38:03.72551452 +0000 UTC m=+810.005390416" Oct 03 16:38:04 crc kubenswrapper[4744]: I1003 16:38:04.668381 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:38:04 crc kubenswrapper[4744]: I1003 16:38:04.669005 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:38:04 crc kubenswrapper[4744]: I1003 16:38:04.669052 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:38:04 crc kubenswrapper[4744]: I1003 16:38:04.669647 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"08b50c5cd69a2f26327ea9dc4a127dffa53b9765528a51178793482f7f49b17f"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 16:38:04 crc kubenswrapper[4744]: I1003 16:38:04.669707 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://08b50c5cd69a2f26327ea9dc4a127dffa53b9765528a51178793482f7f49b17f" gracePeriod=600 Oct 03 16:38:05 crc kubenswrapper[4744]: I1003 16:38:05.716735 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="08b50c5cd69a2f26327ea9dc4a127dffa53b9765528a51178793482f7f49b17f" exitCode=0 Oct 03 16:38:05 crc kubenswrapper[4744]: I1003 16:38:05.716833 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"08b50c5cd69a2f26327ea9dc4a127dffa53b9765528a51178793482f7f49b17f"} Oct 03 16:38:05 crc kubenswrapper[4744]: I1003 16:38:05.717712 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"7ad0acf841a19f4f4edda47a15b6310ad0b8ad88f8fb721a55d7cdc8e8814147"} Oct 03 16:38:05 crc kubenswrapper[4744]: I1003 16:38:05.717764 4744 scope.go:117] "RemoveContainer" containerID="c5619e90872156a7cc8093cc5e7ceeb12da127b0d8d653ccf7ad40c5326f1a97" Oct 03 16:38:06 crc kubenswrapper[4744]: I1003 16:38:06.183908 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:38:06 crc kubenswrapper[4744]: I1003 16:38:06.184399 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:38:06 crc kubenswrapper[4744]: I1003 16:38:06.250158 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:38:06 crc kubenswrapper[4744]: I1003 16:38:06.802842 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:38:07 crc kubenswrapper[4744]: I1003 16:38:07.424895 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r58bx"] Oct 03 16:38:08 crc kubenswrapper[4744]: I1003 16:38:08.745415 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-r58bx" podUID="4c0ed22e-7dce-40f3-8221-d715419dbcfa" containerName="registry-server" containerID="cri-o://c0de36770362e004f318febc7ba3a9330fbfc0577f062503427a0bab88b942bf" gracePeriod=2 Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.148298 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.250977 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.251431 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.281392 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c0ed22e-7dce-40f3-8221-d715419dbcfa-catalog-content\") pod \"4c0ed22e-7dce-40f3-8221-d715419dbcfa\" (UID: \"4c0ed22e-7dce-40f3-8221-d715419dbcfa\") " Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.281540 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c0ed22e-7dce-40f3-8221-d715419dbcfa-utilities\") pod \"4c0ed22e-7dce-40f3-8221-d715419dbcfa\" (UID: \"4c0ed22e-7dce-40f3-8221-d715419dbcfa\") " Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.281580 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fr9n4\" (UniqueName: \"kubernetes.io/projected/4c0ed22e-7dce-40f3-8221-d715419dbcfa-kube-api-access-fr9n4\") pod \"4c0ed22e-7dce-40f3-8221-d715419dbcfa\" (UID: \"4c0ed22e-7dce-40f3-8221-d715419dbcfa\") " Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.283630 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c0ed22e-7dce-40f3-8221-d715419dbcfa-utilities" (OuterVolumeSpecName: "utilities") pod "4c0ed22e-7dce-40f3-8221-d715419dbcfa" (UID: "4c0ed22e-7dce-40f3-8221-d715419dbcfa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.296173 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c0ed22e-7dce-40f3-8221-d715419dbcfa-kube-api-access-fr9n4" (OuterVolumeSpecName: "kube-api-access-fr9n4") pod "4c0ed22e-7dce-40f3-8221-d715419dbcfa" (UID: "4c0ed22e-7dce-40f3-8221-d715419dbcfa"). InnerVolumeSpecName "kube-api-access-fr9n4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.300470 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c0ed22e-7dce-40f3-8221-d715419dbcfa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c0ed22e-7dce-40f3-8221-d715419dbcfa" (UID: "4c0ed22e-7dce-40f3-8221-d715419dbcfa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.330732 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.383647 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c0ed22e-7dce-40f3-8221-d715419dbcfa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.384003 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c0ed22e-7dce-40f3-8221-d715419dbcfa-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.384025 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fr9n4\" (UniqueName: \"kubernetes.io/projected/4c0ed22e-7dce-40f3-8221-d715419dbcfa-kube-api-access-fr9n4\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.755655 4744 generic.go:334] "Generic (PLEG): container finished" podID="4c0ed22e-7dce-40f3-8221-d715419dbcfa" containerID="c0de36770362e004f318febc7ba3a9330fbfc0577f062503427a0bab88b942bf" exitCode=0 Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.755706 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r58bx" event={"ID":"4c0ed22e-7dce-40f3-8221-d715419dbcfa","Type":"ContainerDied","Data":"c0de36770362e004f318febc7ba3a9330fbfc0577f062503427a0bab88b942bf"} Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.756084 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r58bx" event={"ID":"4c0ed22e-7dce-40f3-8221-d715419dbcfa","Type":"ContainerDied","Data":"c7c00cddc025d9cbbf6d1bd85fb284d1e4894bddd2419dbdc446531c91aea7c4"} Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.756125 4744 scope.go:117] "RemoveContainer" containerID="c0de36770362e004f318febc7ba3a9330fbfc0577f062503427a0bab88b942bf" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.755774 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r58bx" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.777795 4744 scope.go:117] "RemoveContainer" containerID="dc5de813cefa3ea24877c079e128145fb002dca798cdcca34355d7a7278e056f" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.807831 4744 scope.go:117] "RemoveContainer" containerID="d769a017abc9b678473d925f95f5916af38403609b44bfc364c8cd6d9eb8d5d7" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.807894 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.829619 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r58bx"] Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.837196 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-r58bx"] Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.839641 4744 scope.go:117] "RemoveContainer" containerID="c0de36770362e004f318febc7ba3a9330fbfc0577f062503427a0bab88b942bf" Oct 03 16:38:09 crc kubenswrapper[4744]: E1003 16:38:09.840257 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0de36770362e004f318febc7ba3a9330fbfc0577f062503427a0bab88b942bf\": container with ID starting with c0de36770362e004f318febc7ba3a9330fbfc0577f062503427a0bab88b942bf not found: ID does not exist" containerID="c0de36770362e004f318febc7ba3a9330fbfc0577f062503427a0bab88b942bf" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.840296 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0de36770362e004f318febc7ba3a9330fbfc0577f062503427a0bab88b942bf"} err="failed to get container status \"c0de36770362e004f318febc7ba3a9330fbfc0577f062503427a0bab88b942bf\": rpc error: code = NotFound desc = could not find container \"c0de36770362e004f318febc7ba3a9330fbfc0577f062503427a0bab88b942bf\": container with ID starting with c0de36770362e004f318febc7ba3a9330fbfc0577f062503427a0bab88b942bf not found: ID does not exist" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.840323 4744 scope.go:117] "RemoveContainer" containerID="dc5de813cefa3ea24877c079e128145fb002dca798cdcca34355d7a7278e056f" Oct 03 16:38:09 crc kubenswrapper[4744]: E1003 16:38:09.840995 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc5de813cefa3ea24877c079e128145fb002dca798cdcca34355d7a7278e056f\": container with ID starting with dc5de813cefa3ea24877c079e128145fb002dca798cdcca34355d7a7278e056f not found: ID does not exist" containerID="dc5de813cefa3ea24877c079e128145fb002dca798cdcca34355d7a7278e056f" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.841060 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc5de813cefa3ea24877c079e128145fb002dca798cdcca34355d7a7278e056f"} err="failed to get container status \"dc5de813cefa3ea24877c079e128145fb002dca798cdcca34355d7a7278e056f\": rpc error: code = NotFound desc = could not find container \"dc5de813cefa3ea24877c079e128145fb002dca798cdcca34355d7a7278e056f\": container with ID starting with dc5de813cefa3ea24877c079e128145fb002dca798cdcca34355d7a7278e056f not found: ID does not exist" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.841102 4744 scope.go:117] "RemoveContainer" containerID="d769a017abc9b678473d925f95f5916af38403609b44bfc364c8cd6d9eb8d5d7" Oct 03 16:38:09 crc kubenswrapper[4744]: E1003 16:38:09.841615 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d769a017abc9b678473d925f95f5916af38403609b44bfc364c8cd6d9eb8d5d7\": container with ID starting with d769a017abc9b678473d925f95f5916af38403609b44bfc364c8cd6d9eb8d5d7 not found: ID does not exist" containerID="d769a017abc9b678473d925f95f5916af38403609b44bfc364c8cd6d9eb8d5d7" Oct 03 16:38:09 crc kubenswrapper[4744]: I1003 16:38:09.841743 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d769a017abc9b678473d925f95f5916af38403609b44bfc364c8cd6d9eb8d5d7"} err="failed to get container status \"d769a017abc9b678473d925f95f5916af38403609b44bfc364c8cd6d9eb8d5d7\": rpc error: code = NotFound desc = could not find container \"d769a017abc9b678473d925f95f5916af38403609b44bfc364c8cd6d9eb8d5d7\": container with ID starting with d769a017abc9b678473d925f95f5916af38403609b44bfc364c8cd6d9eb8d5d7 not found: ID does not exist" Oct 03 16:38:10 crc kubenswrapper[4744]: I1003 16:38:10.902792 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c0ed22e-7dce-40f3-8221-d715419dbcfa" path="/var/lib/kubelet/pods/4c0ed22e-7dce-40f3-8221-d715419dbcfa/volumes" Oct 03 16:38:11 crc kubenswrapper[4744]: I1003 16:38:11.819926 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2k6kz"] Oct 03 16:38:12 crc kubenswrapper[4744]: I1003 16:38:12.727670 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-v5snt" podUID="d61086b0-b78f-4239-8d9a-49aba28c5820" containerName="console" containerID="cri-o://e5a2e9f9e8e08257f04a519e7f0d4d938d3a17733f7b72d5b92b5764268151e4" gracePeriod=15 Oct 03 16:38:12 crc kubenswrapper[4744]: I1003 16:38:12.776106 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2k6kz" podUID="c840ac6c-28ba-4c36-aa15-30dbab14a0d7" containerName="registry-server" containerID="cri-o://1d225d14fd98ae0698cf45e6ab542eb826cc5ee082785de0a049ce816e5d59f0" gracePeriod=2 Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.651885 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-v5snt_d61086b0-b78f-4239-8d9a-49aba28c5820/console/0.log" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.652436 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.711243 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.749925 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-trusted-ca-bundle\") pod \"d61086b0-b78f-4239-8d9a-49aba28c5820\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.750007 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d61086b0-b78f-4239-8d9a-49aba28c5820-console-oauth-config\") pod \"d61086b0-b78f-4239-8d9a-49aba28c5820\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.750150 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jqnn\" (UniqueName: \"kubernetes.io/projected/d61086b0-b78f-4239-8d9a-49aba28c5820-kube-api-access-6jqnn\") pod \"d61086b0-b78f-4239-8d9a-49aba28c5820\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.750195 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d61086b0-b78f-4239-8d9a-49aba28c5820-console-serving-cert\") pod \"d61086b0-b78f-4239-8d9a-49aba28c5820\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.750252 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-console-config\") pod \"d61086b0-b78f-4239-8d9a-49aba28c5820\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.750329 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-service-ca\") pod \"d61086b0-b78f-4239-8d9a-49aba28c5820\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.750410 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-oauth-serving-cert\") pod \"d61086b0-b78f-4239-8d9a-49aba28c5820\" (UID: \"d61086b0-b78f-4239-8d9a-49aba28c5820\") " Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.750843 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "d61086b0-b78f-4239-8d9a-49aba28c5820" (UID: "d61086b0-b78f-4239-8d9a-49aba28c5820"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.752508 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "d61086b0-b78f-4239-8d9a-49aba28c5820" (UID: "d61086b0-b78f-4239-8d9a-49aba28c5820"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.752532 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-service-ca" (OuterVolumeSpecName: "service-ca") pod "d61086b0-b78f-4239-8d9a-49aba28c5820" (UID: "d61086b0-b78f-4239-8d9a-49aba28c5820"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.752725 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-console-config" (OuterVolumeSpecName: "console-config") pod "d61086b0-b78f-4239-8d9a-49aba28c5820" (UID: "d61086b0-b78f-4239-8d9a-49aba28c5820"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.757578 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d61086b0-b78f-4239-8d9a-49aba28c5820-kube-api-access-6jqnn" (OuterVolumeSpecName: "kube-api-access-6jqnn") pod "d61086b0-b78f-4239-8d9a-49aba28c5820" (UID: "d61086b0-b78f-4239-8d9a-49aba28c5820"). InnerVolumeSpecName "kube-api-access-6jqnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.759619 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d61086b0-b78f-4239-8d9a-49aba28c5820-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "d61086b0-b78f-4239-8d9a-49aba28c5820" (UID: "d61086b0-b78f-4239-8d9a-49aba28c5820"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.762885 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d61086b0-b78f-4239-8d9a-49aba28c5820-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "d61086b0-b78f-4239-8d9a-49aba28c5820" (UID: "d61086b0-b78f-4239-8d9a-49aba28c5820"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.786956 4744 generic.go:334] "Generic (PLEG): container finished" podID="c840ac6c-28ba-4c36-aa15-30dbab14a0d7" containerID="1d225d14fd98ae0698cf45e6ab542eb826cc5ee082785de0a049ce816e5d59f0" exitCode=0 Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.787059 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2k6kz" event={"ID":"c840ac6c-28ba-4c36-aa15-30dbab14a0d7","Type":"ContainerDied","Data":"1d225d14fd98ae0698cf45e6ab542eb826cc5ee082785de0a049ce816e5d59f0"} Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.787109 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2k6kz" event={"ID":"c840ac6c-28ba-4c36-aa15-30dbab14a0d7","Type":"ContainerDied","Data":"0827bc2bc1aa41df77d0e130b98d98afa69c47eaf47a3c11470df5cc2d00596f"} Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.787126 4744 scope.go:117] "RemoveContainer" containerID="1d225d14fd98ae0698cf45e6ab542eb826cc5ee082785de0a049ce816e5d59f0" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.787389 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2k6kz" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.789286 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-v5snt_d61086b0-b78f-4239-8d9a-49aba28c5820/console/0.log" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.789314 4744 generic.go:334] "Generic (PLEG): container finished" podID="d61086b0-b78f-4239-8d9a-49aba28c5820" containerID="e5a2e9f9e8e08257f04a519e7f0d4d938d3a17733f7b72d5b92b5764268151e4" exitCode=2 Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.789331 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-v5snt" event={"ID":"d61086b0-b78f-4239-8d9a-49aba28c5820","Type":"ContainerDied","Data":"e5a2e9f9e8e08257f04a519e7f0d4d938d3a17733f7b72d5b92b5764268151e4"} Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.789344 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-v5snt" event={"ID":"d61086b0-b78f-4239-8d9a-49aba28c5820","Type":"ContainerDied","Data":"ab1fb08e373c4fe5a018e7df29d9d99216c0be26b95d9127c213773e2d982cb6"} Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.789386 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-v5snt" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.805036 4744 scope.go:117] "RemoveContainer" containerID="1e4e882ebe160e3c898d682ac297051575b4f051d09371fee97dc07d542fb609" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.817207 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-v5snt"] Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.821620 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-v5snt"] Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.831769 4744 scope.go:117] "RemoveContainer" containerID="32536b28176c56f4dce2657605ef79a68be4e83f283e551eaead23d8f7ad0cb5" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.853421 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckhcb\" (UniqueName: \"kubernetes.io/projected/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-kube-api-access-ckhcb\") pod \"c840ac6c-28ba-4c36-aa15-30dbab14a0d7\" (UID: \"c840ac6c-28ba-4c36-aa15-30dbab14a0d7\") " Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.853903 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-utilities\") pod \"c840ac6c-28ba-4c36-aa15-30dbab14a0d7\" (UID: \"c840ac6c-28ba-4c36-aa15-30dbab14a0d7\") " Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.854006 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-catalog-content\") pod \"c840ac6c-28ba-4c36-aa15-30dbab14a0d7\" (UID: \"c840ac6c-28ba-4c36-aa15-30dbab14a0d7\") " Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.854536 4744 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.854566 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.854586 4744 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d61086b0-b78f-4239-8d9a-49aba28c5820-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.854604 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jqnn\" (UniqueName: \"kubernetes.io/projected/d61086b0-b78f-4239-8d9a-49aba28c5820-kube-api-access-6jqnn\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.854622 4744 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d61086b0-b78f-4239-8d9a-49aba28c5820-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.854641 4744 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-console-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.854659 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d61086b0-b78f-4239-8d9a-49aba28c5820-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.854691 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-utilities" (OuterVolumeSpecName: "utilities") pod "c840ac6c-28ba-4c36-aa15-30dbab14a0d7" (UID: "c840ac6c-28ba-4c36-aa15-30dbab14a0d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.858205 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-kube-api-access-ckhcb" (OuterVolumeSpecName: "kube-api-access-ckhcb") pod "c840ac6c-28ba-4c36-aa15-30dbab14a0d7" (UID: "c840ac6c-28ba-4c36-aa15-30dbab14a0d7"). InnerVolumeSpecName "kube-api-access-ckhcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.865608 4744 scope.go:117] "RemoveContainer" containerID="1d225d14fd98ae0698cf45e6ab542eb826cc5ee082785de0a049ce816e5d59f0" Oct 03 16:38:13 crc kubenswrapper[4744]: E1003 16:38:13.866077 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d225d14fd98ae0698cf45e6ab542eb826cc5ee082785de0a049ce816e5d59f0\": container with ID starting with 1d225d14fd98ae0698cf45e6ab542eb826cc5ee082785de0a049ce816e5d59f0 not found: ID does not exist" containerID="1d225d14fd98ae0698cf45e6ab542eb826cc5ee082785de0a049ce816e5d59f0" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.866119 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d225d14fd98ae0698cf45e6ab542eb826cc5ee082785de0a049ce816e5d59f0"} err="failed to get container status \"1d225d14fd98ae0698cf45e6ab542eb826cc5ee082785de0a049ce816e5d59f0\": rpc error: code = NotFound desc = could not find container \"1d225d14fd98ae0698cf45e6ab542eb826cc5ee082785de0a049ce816e5d59f0\": container with ID starting with 1d225d14fd98ae0698cf45e6ab542eb826cc5ee082785de0a049ce816e5d59f0 not found: ID does not exist" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.866147 4744 scope.go:117] "RemoveContainer" containerID="1e4e882ebe160e3c898d682ac297051575b4f051d09371fee97dc07d542fb609" Oct 03 16:38:13 crc kubenswrapper[4744]: E1003 16:38:13.866469 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e4e882ebe160e3c898d682ac297051575b4f051d09371fee97dc07d542fb609\": container with ID starting with 1e4e882ebe160e3c898d682ac297051575b4f051d09371fee97dc07d542fb609 not found: ID does not exist" containerID="1e4e882ebe160e3c898d682ac297051575b4f051d09371fee97dc07d542fb609" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.866563 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e4e882ebe160e3c898d682ac297051575b4f051d09371fee97dc07d542fb609"} err="failed to get container status \"1e4e882ebe160e3c898d682ac297051575b4f051d09371fee97dc07d542fb609\": rpc error: code = NotFound desc = could not find container \"1e4e882ebe160e3c898d682ac297051575b4f051d09371fee97dc07d542fb609\": container with ID starting with 1e4e882ebe160e3c898d682ac297051575b4f051d09371fee97dc07d542fb609 not found: ID does not exist" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.866617 4744 scope.go:117] "RemoveContainer" containerID="32536b28176c56f4dce2657605ef79a68be4e83f283e551eaead23d8f7ad0cb5" Oct 03 16:38:13 crc kubenswrapper[4744]: E1003 16:38:13.867129 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32536b28176c56f4dce2657605ef79a68be4e83f283e551eaead23d8f7ad0cb5\": container with ID starting with 32536b28176c56f4dce2657605ef79a68be4e83f283e551eaead23d8f7ad0cb5 not found: ID does not exist" containerID="32536b28176c56f4dce2657605ef79a68be4e83f283e551eaead23d8f7ad0cb5" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.867153 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32536b28176c56f4dce2657605ef79a68be4e83f283e551eaead23d8f7ad0cb5"} err="failed to get container status \"32536b28176c56f4dce2657605ef79a68be4e83f283e551eaead23d8f7ad0cb5\": rpc error: code = NotFound desc = could not find container \"32536b28176c56f4dce2657605ef79a68be4e83f283e551eaead23d8f7ad0cb5\": container with ID starting with 32536b28176c56f4dce2657605ef79a68be4e83f283e551eaead23d8f7ad0cb5 not found: ID does not exist" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.867169 4744 scope.go:117] "RemoveContainer" containerID="e5a2e9f9e8e08257f04a519e7f0d4d938d3a17733f7b72d5b92b5764268151e4" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.898102 4744 scope.go:117] "RemoveContainer" containerID="e5a2e9f9e8e08257f04a519e7f0d4d938d3a17733f7b72d5b92b5764268151e4" Oct 03 16:38:13 crc kubenswrapper[4744]: E1003 16:38:13.898534 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5a2e9f9e8e08257f04a519e7f0d4d938d3a17733f7b72d5b92b5764268151e4\": container with ID starting with e5a2e9f9e8e08257f04a519e7f0d4d938d3a17733f7b72d5b92b5764268151e4 not found: ID does not exist" containerID="e5a2e9f9e8e08257f04a519e7f0d4d938d3a17733f7b72d5b92b5764268151e4" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.898589 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5a2e9f9e8e08257f04a519e7f0d4d938d3a17733f7b72d5b92b5764268151e4"} err="failed to get container status \"e5a2e9f9e8e08257f04a519e7f0d4d938d3a17733f7b72d5b92b5764268151e4\": rpc error: code = NotFound desc = could not find container \"e5a2e9f9e8e08257f04a519e7f0d4d938d3a17733f7b72d5b92b5764268151e4\": container with ID starting with e5a2e9f9e8e08257f04a519e7f0d4d938d3a17733f7b72d5b92b5764268151e4 not found: ID does not exist" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.902859 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c840ac6c-28ba-4c36-aa15-30dbab14a0d7" (UID: "c840ac6c-28ba-4c36-aa15-30dbab14a0d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.956450 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.956498 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:13 crc kubenswrapper[4744]: I1003 16:38:13.956530 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckhcb\" (UniqueName: \"kubernetes.io/projected/c840ac6c-28ba-4c36-aa15-30dbab14a0d7-kube-api-access-ckhcb\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:14 crc kubenswrapper[4744]: I1003 16:38:14.118821 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2k6kz"] Oct 03 16:38:14 crc kubenswrapper[4744]: I1003 16:38:14.125824 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2k6kz"] Oct 03 16:38:14 crc kubenswrapper[4744]: I1003 16:38:14.904694 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c840ac6c-28ba-4c36-aa15-30dbab14a0d7" path="/var/lib/kubelet/pods/c840ac6c-28ba-4c36-aa15-30dbab14a0d7/volumes" Oct 03 16:38:14 crc kubenswrapper[4744]: I1003 16:38:14.906525 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d61086b0-b78f-4239-8d9a-49aba28c5820" path="/var/lib/kubelet/pods/d61086b0-b78f-4239-8d9a-49aba28c5820/volumes" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.483545 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh"] Oct 03 16:38:18 crc kubenswrapper[4744]: E1003 16:38:18.483813 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c840ac6c-28ba-4c36-aa15-30dbab14a0d7" containerName="registry-server" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.483829 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c840ac6c-28ba-4c36-aa15-30dbab14a0d7" containerName="registry-server" Oct 03 16:38:18 crc kubenswrapper[4744]: E1003 16:38:18.483840 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c0ed22e-7dce-40f3-8221-d715419dbcfa" containerName="extract-utilities" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.483848 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c0ed22e-7dce-40f3-8221-d715419dbcfa" containerName="extract-utilities" Oct 03 16:38:18 crc kubenswrapper[4744]: E1003 16:38:18.483858 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c0ed22e-7dce-40f3-8221-d715419dbcfa" containerName="registry-server" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.483867 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c0ed22e-7dce-40f3-8221-d715419dbcfa" containerName="registry-server" Oct 03 16:38:18 crc kubenswrapper[4744]: E1003 16:38:18.483876 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c840ac6c-28ba-4c36-aa15-30dbab14a0d7" containerName="extract-content" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.483884 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c840ac6c-28ba-4c36-aa15-30dbab14a0d7" containerName="extract-content" Oct 03 16:38:18 crc kubenswrapper[4744]: E1003 16:38:18.483898 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d61086b0-b78f-4239-8d9a-49aba28c5820" containerName="console" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.483906 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d61086b0-b78f-4239-8d9a-49aba28c5820" containerName="console" Oct 03 16:38:18 crc kubenswrapper[4744]: E1003 16:38:18.483921 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c0ed22e-7dce-40f3-8221-d715419dbcfa" containerName="extract-content" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.483928 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c0ed22e-7dce-40f3-8221-d715419dbcfa" containerName="extract-content" Oct 03 16:38:18 crc kubenswrapper[4744]: E1003 16:38:18.483943 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c840ac6c-28ba-4c36-aa15-30dbab14a0d7" containerName="extract-utilities" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.483951 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c840ac6c-28ba-4c36-aa15-30dbab14a0d7" containerName="extract-utilities" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.484074 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c0ed22e-7dce-40f3-8221-d715419dbcfa" containerName="registry-server" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.484088 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c840ac6c-28ba-4c36-aa15-30dbab14a0d7" containerName="registry-server" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.484102 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d61086b0-b78f-4239-8d9a-49aba28c5820" containerName="console" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.485180 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.487836 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.496672 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh"] Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.628760 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c795770c-99b7-4077-a8b4-f50b678f32b8-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh\" (UID: \"c795770c-99b7-4077-a8b4-f50b678f32b8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.629169 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll6tz\" (UniqueName: \"kubernetes.io/projected/c795770c-99b7-4077-a8b4-f50b678f32b8-kube-api-access-ll6tz\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh\" (UID: \"c795770c-99b7-4077-a8b4-f50b678f32b8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.629275 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c795770c-99b7-4077-a8b4-f50b678f32b8-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh\" (UID: \"c795770c-99b7-4077-a8b4-f50b678f32b8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.730865 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c795770c-99b7-4077-a8b4-f50b678f32b8-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh\" (UID: \"c795770c-99b7-4077-a8b4-f50b678f32b8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.731036 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll6tz\" (UniqueName: \"kubernetes.io/projected/c795770c-99b7-4077-a8b4-f50b678f32b8-kube-api-access-ll6tz\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh\" (UID: \"c795770c-99b7-4077-a8b4-f50b678f32b8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.731079 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c795770c-99b7-4077-a8b4-f50b678f32b8-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh\" (UID: \"c795770c-99b7-4077-a8b4-f50b678f32b8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.731721 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c795770c-99b7-4077-a8b4-f50b678f32b8-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh\" (UID: \"c795770c-99b7-4077-a8b4-f50b678f32b8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.731768 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c795770c-99b7-4077-a8b4-f50b678f32b8-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh\" (UID: \"c795770c-99b7-4077-a8b4-f50b678f32b8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.770783 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll6tz\" (UniqueName: \"kubernetes.io/projected/c795770c-99b7-4077-a8b4-f50b678f32b8-kube-api-access-ll6tz\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh\" (UID: \"c795770c-99b7-4077-a8b4-f50b678f32b8\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" Oct 03 16:38:18 crc kubenswrapper[4744]: I1003 16:38:18.811842 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" Oct 03 16:38:19 crc kubenswrapper[4744]: I1003 16:38:19.251116 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh"] Oct 03 16:38:19 crc kubenswrapper[4744]: I1003 16:38:19.839821 4744 generic.go:334] "Generic (PLEG): container finished" podID="c795770c-99b7-4077-a8b4-f50b678f32b8" containerID="df00eeaa0622cf8c0330703cf3e0779f8bddb63236b526457ddaffcb41ee79ff" exitCode=0 Oct 03 16:38:19 crc kubenswrapper[4744]: I1003 16:38:19.839876 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" event={"ID":"c795770c-99b7-4077-a8b4-f50b678f32b8","Type":"ContainerDied","Data":"df00eeaa0622cf8c0330703cf3e0779f8bddb63236b526457ddaffcb41ee79ff"} Oct 03 16:38:19 crc kubenswrapper[4744]: I1003 16:38:19.839909 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" event={"ID":"c795770c-99b7-4077-a8b4-f50b678f32b8","Type":"ContainerStarted","Data":"f76da7b7edeee4c35191cd9d099ef9f1abac0754dcb97aec4453b4e90bab9ac7"} Oct 03 16:38:23 crc kubenswrapper[4744]: I1003 16:38:23.876321 4744 generic.go:334] "Generic (PLEG): container finished" podID="c795770c-99b7-4077-a8b4-f50b678f32b8" containerID="11f76fce6f8ee7408bc260760dfff899231e5dd0b085e318d72947596ae6a4a9" exitCode=0 Oct 03 16:38:23 crc kubenswrapper[4744]: I1003 16:38:23.876415 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" event={"ID":"c795770c-99b7-4077-a8b4-f50b678f32b8","Type":"ContainerDied","Data":"11f76fce6f8ee7408bc260760dfff899231e5dd0b085e318d72947596ae6a4a9"} Oct 03 16:38:24 crc kubenswrapper[4744]: I1003 16:38:24.885778 4744 generic.go:334] "Generic (PLEG): container finished" podID="c795770c-99b7-4077-a8b4-f50b678f32b8" containerID="fa002dab00f54e40babf39ed80032a5590fc3c28bdd59a095a97653b560dee26" exitCode=0 Oct 03 16:38:24 crc kubenswrapper[4744]: I1003 16:38:24.885828 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" event={"ID":"c795770c-99b7-4077-a8b4-f50b678f32b8","Type":"ContainerDied","Data":"fa002dab00f54e40babf39ed80032a5590fc3c28bdd59a095a97653b560dee26"} Oct 03 16:38:26 crc kubenswrapper[4744]: I1003 16:38:26.225288 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" Oct 03 16:38:26 crc kubenswrapper[4744]: I1003 16:38:26.345332 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c795770c-99b7-4077-a8b4-f50b678f32b8-bundle\") pod \"c795770c-99b7-4077-a8b4-f50b678f32b8\" (UID: \"c795770c-99b7-4077-a8b4-f50b678f32b8\") " Oct 03 16:38:26 crc kubenswrapper[4744]: I1003 16:38:26.345902 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c795770c-99b7-4077-a8b4-f50b678f32b8-util\") pod \"c795770c-99b7-4077-a8b4-f50b678f32b8\" (UID: \"c795770c-99b7-4077-a8b4-f50b678f32b8\") " Oct 03 16:38:26 crc kubenswrapper[4744]: I1003 16:38:26.345974 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ll6tz\" (UniqueName: \"kubernetes.io/projected/c795770c-99b7-4077-a8b4-f50b678f32b8-kube-api-access-ll6tz\") pod \"c795770c-99b7-4077-a8b4-f50b678f32b8\" (UID: \"c795770c-99b7-4077-a8b4-f50b678f32b8\") " Oct 03 16:38:26 crc kubenswrapper[4744]: I1003 16:38:26.346970 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c795770c-99b7-4077-a8b4-f50b678f32b8-bundle" (OuterVolumeSpecName: "bundle") pod "c795770c-99b7-4077-a8b4-f50b678f32b8" (UID: "c795770c-99b7-4077-a8b4-f50b678f32b8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:38:26 crc kubenswrapper[4744]: I1003 16:38:26.355352 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c795770c-99b7-4077-a8b4-f50b678f32b8-kube-api-access-ll6tz" (OuterVolumeSpecName: "kube-api-access-ll6tz") pod "c795770c-99b7-4077-a8b4-f50b678f32b8" (UID: "c795770c-99b7-4077-a8b4-f50b678f32b8"). InnerVolumeSpecName "kube-api-access-ll6tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:38:26 crc kubenswrapper[4744]: I1003 16:38:26.364171 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c795770c-99b7-4077-a8b4-f50b678f32b8-util" (OuterVolumeSpecName: "util") pod "c795770c-99b7-4077-a8b4-f50b678f32b8" (UID: "c795770c-99b7-4077-a8b4-f50b678f32b8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:38:26 crc kubenswrapper[4744]: I1003 16:38:26.448478 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c795770c-99b7-4077-a8b4-f50b678f32b8-util\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:26 crc kubenswrapper[4744]: I1003 16:38:26.448531 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ll6tz\" (UniqueName: \"kubernetes.io/projected/c795770c-99b7-4077-a8b4-f50b678f32b8-kube-api-access-ll6tz\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:26 crc kubenswrapper[4744]: I1003 16:38:26.448542 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c795770c-99b7-4077-a8b4-f50b678f32b8-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:38:26 crc kubenswrapper[4744]: I1003 16:38:26.905835 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" event={"ID":"c795770c-99b7-4077-a8b4-f50b678f32b8","Type":"ContainerDied","Data":"f76da7b7edeee4c35191cd9d099ef9f1abac0754dcb97aec4453b4e90bab9ac7"} Oct 03 16:38:26 crc kubenswrapper[4744]: I1003 16:38:26.905888 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f76da7b7edeee4c35191cd9d099ef9f1abac0754dcb97aec4453b4e90bab9ac7" Oct 03 16:38:26 crc kubenswrapper[4744]: I1003 16:38:26.905891 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.673646 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8"] Oct 03 16:38:35 crc kubenswrapper[4744]: E1003 16:38:35.674545 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c795770c-99b7-4077-a8b4-f50b678f32b8" containerName="util" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.674562 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c795770c-99b7-4077-a8b4-f50b678f32b8" containerName="util" Oct 03 16:38:35 crc kubenswrapper[4744]: E1003 16:38:35.674578 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c795770c-99b7-4077-a8b4-f50b678f32b8" containerName="pull" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.674586 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c795770c-99b7-4077-a8b4-f50b678f32b8" containerName="pull" Oct 03 16:38:35 crc kubenswrapper[4744]: E1003 16:38:35.674598 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c795770c-99b7-4077-a8b4-f50b678f32b8" containerName="extract" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.674608 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c795770c-99b7-4077-a8b4-f50b678f32b8" containerName="extract" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.674749 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c795770c-99b7-4077-a8b4-f50b678f32b8" containerName="extract" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.675247 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.678284 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.678471 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.678538 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.678318 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.678871 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-l4dhn" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.695367 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8"] Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.788920 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9mqk\" (UniqueName: \"kubernetes.io/projected/8776a23e-e0c0-4c8d-84e8-d34d5036e3c6-kube-api-access-x9mqk\") pod \"metallb-operator-controller-manager-6cbff679b9-cdsl8\" (UID: \"8776a23e-e0c0-4c8d-84e8-d34d5036e3c6\") " pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.789201 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8776a23e-e0c0-4c8d-84e8-d34d5036e3c6-webhook-cert\") pod \"metallb-operator-controller-manager-6cbff679b9-cdsl8\" (UID: \"8776a23e-e0c0-4c8d-84e8-d34d5036e3c6\") " pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.789313 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8776a23e-e0c0-4c8d-84e8-d34d5036e3c6-apiservice-cert\") pod \"metallb-operator-controller-manager-6cbff679b9-cdsl8\" (UID: \"8776a23e-e0c0-4c8d-84e8-d34d5036e3c6\") " pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.890198 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8776a23e-e0c0-4c8d-84e8-d34d5036e3c6-webhook-cert\") pod \"metallb-operator-controller-manager-6cbff679b9-cdsl8\" (UID: \"8776a23e-e0c0-4c8d-84e8-d34d5036e3c6\") " pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.890270 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8776a23e-e0c0-4c8d-84e8-d34d5036e3c6-apiservice-cert\") pod \"metallb-operator-controller-manager-6cbff679b9-cdsl8\" (UID: \"8776a23e-e0c0-4c8d-84e8-d34d5036e3c6\") " pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.890308 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9mqk\" (UniqueName: \"kubernetes.io/projected/8776a23e-e0c0-4c8d-84e8-d34d5036e3c6-kube-api-access-x9mqk\") pod \"metallb-operator-controller-manager-6cbff679b9-cdsl8\" (UID: \"8776a23e-e0c0-4c8d-84e8-d34d5036e3c6\") " pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.899194 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8776a23e-e0c0-4c8d-84e8-d34d5036e3c6-apiservice-cert\") pod \"metallb-operator-controller-manager-6cbff679b9-cdsl8\" (UID: \"8776a23e-e0c0-4c8d-84e8-d34d5036e3c6\") " pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.899338 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8776a23e-e0c0-4c8d-84e8-d34d5036e3c6-webhook-cert\") pod \"metallb-operator-controller-manager-6cbff679b9-cdsl8\" (UID: \"8776a23e-e0c0-4c8d-84e8-d34d5036e3c6\") " pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.921875 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg"] Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.922477 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9mqk\" (UniqueName: \"kubernetes.io/projected/8776a23e-e0c0-4c8d-84e8-d34d5036e3c6-kube-api-access-x9mqk\") pod \"metallb-operator-controller-manager-6cbff679b9-cdsl8\" (UID: \"8776a23e-e0c0-4c8d-84e8-d34d5036e3c6\") " pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.922640 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.924708 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.925051 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.925274 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-kw4nr" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.940516 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg"] Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.991157 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.991636 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/259f45bb-e4ab-45f4-9d77-5e788361c35a-apiservice-cert\") pod \"metallb-operator-webhook-server-86c9f64ddf-r6kfg\" (UID: \"259f45bb-e4ab-45f4-9d77-5e788361c35a\") " pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.991700 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/259f45bb-e4ab-45f4-9d77-5e788361c35a-webhook-cert\") pod \"metallb-operator-webhook-server-86c9f64ddf-r6kfg\" (UID: \"259f45bb-e4ab-45f4-9d77-5e788361c35a\") " pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" Oct 03 16:38:35 crc kubenswrapper[4744]: I1003 16:38:35.991770 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsshq\" (UniqueName: \"kubernetes.io/projected/259f45bb-e4ab-45f4-9d77-5e788361c35a-kube-api-access-wsshq\") pod \"metallb-operator-webhook-server-86c9f64ddf-r6kfg\" (UID: \"259f45bb-e4ab-45f4-9d77-5e788361c35a\") " pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" Oct 03 16:38:36 crc kubenswrapper[4744]: I1003 16:38:36.093435 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/259f45bb-e4ab-45f4-9d77-5e788361c35a-apiservice-cert\") pod \"metallb-operator-webhook-server-86c9f64ddf-r6kfg\" (UID: \"259f45bb-e4ab-45f4-9d77-5e788361c35a\") " pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" Oct 03 16:38:36 crc kubenswrapper[4744]: I1003 16:38:36.093709 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/259f45bb-e4ab-45f4-9d77-5e788361c35a-webhook-cert\") pod \"metallb-operator-webhook-server-86c9f64ddf-r6kfg\" (UID: \"259f45bb-e4ab-45f4-9d77-5e788361c35a\") " pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" Oct 03 16:38:36 crc kubenswrapper[4744]: I1003 16:38:36.094010 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsshq\" (UniqueName: \"kubernetes.io/projected/259f45bb-e4ab-45f4-9d77-5e788361c35a-kube-api-access-wsshq\") pod \"metallb-operator-webhook-server-86c9f64ddf-r6kfg\" (UID: \"259f45bb-e4ab-45f4-9d77-5e788361c35a\") " pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" Oct 03 16:38:36 crc kubenswrapper[4744]: I1003 16:38:36.097753 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/259f45bb-e4ab-45f4-9d77-5e788361c35a-webhook-cert\") pod \"metallb-operator-webhook-server-86c9f64ddf-r6kfg\" (UID: \"259f45bb-e4ab-45f4-9d77-5e788361c35a\") " pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" Oct 03 16:38:36 crc kubenswrapper[4744]: I1003 16:38:36.097775 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/259f45bb-e4ab-45f4-9d77-5e788361c35a-apiservice-cert\") pod \"metallb-operator-webhook-server-86c9f64ddf-r6kfg\" (UID: \"259f45bb-e4ab-45f4-9d77-5e788361c35a\") " pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" Oct 03 16:38:36 crc kubenswrapper[4744]: I1003 16:38:36.113713 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsshq\" (UniqueName: \"kubernetes.io/projected/259f45bb-e4ab-45f4-9d77-5e788361c35a-kube-api-access-wsshq\") pod \"metallb-operator-webhook-server-86c9f64ddf-r6kfg\" (UID: \"259f45bb-e4ab-45f4-9d77-5e788361c35a\") " pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" Oct 03 16:38:36 crc kubenswrapper[4744]: I1003 16:38:36.280634 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" Oct 03 16:38:36 crc kubenswrapper[4744]: I1003 16:38:36.516158 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8"] Oct 03 16:38:36 crc kubenswrapper[4744]: W1003 16:38:36.521238 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8776a23e_e0c0_4c8d_84e8_d34d5036e3c6.slice/crio-44324aae31eebc1da9de2efd8fdabeca0a40554037967f566200197aa81da7a8 WatchSource:0}: Error finding container 44324aae31eebc1da9de2efd8fdabeca0a40554037967f566200197aa81da7a8: Status 404 returned error can't find the container with id 44324aae31eebc1da9de2efd8fdabeca0a40554037967f566200197aa81da7a8 Oct 03 16:38:36 crc kubenswrapper[4744]: I1003 16:38:36.821084 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg"] Oct 03 16:38:36 crc kubenswrapper[4744]: W1003 16:38:36.828161 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod259f45bb_e4ab_45f4_9d77_5e788361c35a.slice/crio-4f059341f6b19d5a619ccf5ce7c5c212dbe5979657c9012fa72f82c3c5c7f371 WatchSource:0}: Error finding container 4f059341f6b19d5a619ccf5ce7c5c212dbe5979657c9012fa72f82c3c5c7f371: Status 404 returned error can't find the container with id 4f059341f6b19d5a619ccf5ce7c5c212dbe5979657c9012fa72f82c3c5c7f371 Oct 03 16:38:36 crc kubenswrapper[4744]: I1003 16:38:36.974593 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" event={"ID":"8776a23e-e0c0-4c8d-84e8-d34d5036e3c6","Type":"ContainerStarted","Data":"44324aae31eebc1da9de2efd8fdabeca0a40554037967f566200197aa81da7a8"} Oct 03 16:38:36 crc kubenswrapper[4744]: I1003 16:38:36.976426 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" event={"ID":"259f45bb-e4ab-45f4-9d77-5e788361c35a","Type":"ContainerStarted","Data":"4f059341f6b19d5a619ccf5ce7c5c212dbe5979657c9012fa72f82c3c5c7f371"} Oct 03 16:38:42 crc kubenswrapper[4744]: I1003 16:38:42.010107 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" event={"ID":"8776a23e-e0c0-4c8d-84e8-d34d5036e3c6","Type":"ContainerStarted","Data":"d5469f42cb511d5ad7f2565685ec10b877facf3e2afa218cdea1eca37747426c"} Oct 03 16:38:42 crc kubenswrapper[4744]: I1003 16:38:42.011059 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" Oct 03 16:38:42 crc kubenswrapper[4744]: I1003 16:38:42.012147 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" event={"ID":"259f45bb-e4ab-45f4-9d77-5e788361c35a","Type":"ContainerStarted","Data":"aa7c9d78ea9021ccfdc561ef17c4c1314a9fcf65c586d52d80989d720d33d902"} Oct 03 16:38:42 crc kubenswrapper[4744]: I1003 16:38:42.012612 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" Oct 03 16:38:42 crc kubenswrapper[4744]: I1003 16:38:42.052904 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" podStartSLOduration=2.303625501 podStartE2EDuration="7.05288261s" podCreationTimestamp="2025-10-03 16:38:35 +0000 UTC" firstStartedPulling="2025-10-03 16:38:36.830827184 +0000 UTC m=+843.110703080" lastFinishedPulling="2025-10-03 16:38:41.580084293 +0000 UTC m=+847.859960189" observedRunningTime="2025-10-03 16:38:42.049593155 +0000 UTC m=+848.329469051" watchObservedRunningTime="2025-10-03 16:38:42.05288261 +0000 UTC m=+848.332758506" Oct 03 16:38:42 crc kubenswrapper[4744]: I1003 16:38:42.054336 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" podStartSLOduration=2.01911878 podStartE2EDuration="7.054326537s" podCreationTimestamp="2025-10-03 16:38:35 +0000 UTC" firstStartedPulling="2025-10-03 16:38:36.523931988 +0000 UTC m=+842.803807884" lastFinishedPulling="2025-10-03 16:38:41.559139745 +0000 UTC m=+847.839015641" observedRunningTime="2025-10-03 16:38:42.030771652 +0000 UTC m=+848.310647548" watchObservedRunningTime="2025-10-03 16:38:42.054326537 +0000 UTC m=+848.334202433" Oct 03 16:38:56 crc kubenswrapper[4744]: I1003 16:38:56.287013 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-86c9f64ddf-r6kfg" Oct 03 16:39:15 crc kubenswrapper[4744]: I1003 16:39:15.994726 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6cbff679b9-cdsl8" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.696533 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f"] Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.697484 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.699269 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.699993 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-qx7w4" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.709559 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-7dg9g"] Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.713046 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.716188 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f"] Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.716750 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.716826 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.793233 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-vhk6s"] Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.794171 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56173127-d55d-4c05-ac7b-1cdc4484f3bb-metrics-certs\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.794275 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-vhk6s" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.794359 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vc5m\" (UniqueName: \"kubernetes.io/projected/56173127-d55d-4c05-ac7b-1cdc4484f3bb-kube-api-access-2vc5m\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.794460 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dfgp\" (UniqueName: \"kubernetes.io/projected/703cac15-6674-4529-9a6b-75f646a84d3e-kube-api-access-2dfgp\") pod \"frr-k8s-webhook-server-64bf5d555-xsl5f\" (UID: \"703cac15-6674-4529-9a6b-75f646a84d3e\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.794599 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/56173127-d55d-4c05-ac7b-1cdc4484f3bb-frr-sockets\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.794732 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/56173127-d55d-4c05-ac7b-1cdc4484f3bb-metrics\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.794816 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/56173127-d55d-4c05-ac7b-1cdc4484f3bb-frr-conf\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.794908 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/703cac15-6674-4529-9a6b-75f646a84d3e-cert\") pod \"frr-k8s-webhook-server-64bf5d555-xsl5f\" (UID: \"703cac15-6674-4529-9a6b-75f646a84d3e\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.795014 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/56173127-d55d-4c05-ac7b-1cdc4484f3bb-frr-startup\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.795122 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/56173127-d55d-4c05-ac7b-1cdc4484f3bb-reloader\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: W1003 16:39:16.802615 4744 reflector.go:561] object-"metallb-system"/"controller-certs-secret": failed to list *v1.Secret: secrets "controller-certs-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 03 16:39:16 crc kubenswrapper[4744]: E1003 16:39:16.802660 4744 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"controller-certs-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"controller-certs-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.813790 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-5qx66"] Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.814820 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5qx66" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.820120 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-bfmdf" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.820323 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.820471 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.820628 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.843431 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-vhk6s"] Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896311 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0d474bc7-3b94-40d8-91d6-db976b2975d7-metrics-certs\") pod \"speaker-5qx66\" (UID: \"0d474bc7-3b94-40d8-91d6-db976b2975d7\") " pod="metallb-system/speaker-5qx66" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896383 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/56173127-d55d-4c05-ac7b-1cdc4484f3bb-reloader\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896422 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56173127-d55d-4c05-ac7b-1cdc4484f3bb-metrics-certs\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896462 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vc5m\" (UniqueName: \"kubernetes.io/projected/56173127-d55d-4c05-ac7b-1cdc4484f3bb-kube-api-access-2vc5m\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896512 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dfgp\" (UniqueName: \"kubernetes.io/projected/703cac15-6674-4529-9a6b-75f646a84d3e-kube-api-access-2dfgp\") pod \"frr-k8s-webhook-server-64bf5d555-xsl5f\" (UID: \"703cac15-6674-4529-9a6b-75f646a84d3e\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896534 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/56173127-d55d-4c05-ac7b-1cdc4484f3bb-frr-sockets\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896559 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0d474bc7-3b94-40d8-91d6-db976b2975d7-memberlist\") pod \"speaker-5qx66\" (UID: \"0d474bc7-3b94-40d8-91d6-db976b2975d7\") " pod="metallb-system/speaker-5qx66" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896606 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f04d600-8bb7-4420-aa5f-c1b51a0a4e16-metrics-certs\") pod \"controller-68d546b9d8-vhk6s\" (UID: \"5f04d600-8bb7-4420-aa5f-c1b51a0a4e16\") " pod="metallb-system/controller-68d546b9d8-vhk6s" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896644 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/56173127-d55d-4c05-ac7b-1cdc4484f3bb-metrics\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896670 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/56173127-d55d-4c05-ac7b-1cdc4484f3bb-frr-conf\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896697 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d4w7\" (UniqueName: \"kubernetes.io/projected/5f04d600-8bb7-4420-aa5f-c1b51a0a4e16-kube-api-access-6d4w7\") pod \"controller-68d546b9d8-vhk6s\" (UID: \"5f04d600-8bb7-4420-aa5f-c1b51a0a4e16\") " pod="metallb-system/controller-68d546b9d8-vhk6s" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896724 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbgxb\" (UniqueName: \"kubernetes.io/projected/0d474bc7-3b94-40d8-91d6-db976b2975d7-kube-api-access-lbgxb\") pod \"speaker-5qx66\" (UID: \"0d474bc7-3b94-40d8-91d6-db976b2975d7\") " pod="metallb-system/speaker-5qx66" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896755 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5f04d600-8bb7-4420-aa5f-c1b51a0a4e16-cert\") pod \"controller-68d546b9d8-vhk6s\" (UID: \"5f04d600-8bb7-4420-aa5f-c1b51a0a4e16\") " pod="metallb-system/controller-68d546b9d8-vhk6s" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896782 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/703cac15-6674-4529-9a6b-75f646a84d3e-cert\") pod \"frr-k8s-webhook-server-64bf5d555-xsl5f\" (UID: \"703cac15-6674-4529-9a6b-75f646a84d3e\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896807 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/0d474bc7-3b94-40d8-91d6-db976b2975d7-metallb-excludel2\") pod \"speaker-5qx66\" (UID: \"0d474bc7-3b94-40d8-91d6-db976b2975d7\") " pod="metallb-system/speaker-5qx66" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.896836 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/56173127-d55d-4c05-ac7b-1cdc4484f3bb-frr-startup\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.897829 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/56173127-d55d-4c05-ac7b-1cdc4484f3bb-reloader\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.898023 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/56173127-d55d-4c05-ac7b-1cdc4484f3bb-frr-startup\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.898259 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/56173127-d55d-4c05-ac7b-1cdc4484f3bb-metrics\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.898454 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/56173127-d55d-4c05-ac7b-1cdc4484f3bb-frr-conf\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.899697 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/56173127-d55d-4c05-ac7b-1cdc4484f3bb-frr-sockets\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.909304 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56173127-d55d-4c05-ac7b-1cdc4484f3bb-metrics-certs\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.918157 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/703cac15-6674-4529-9a6b-75f646a84d3e-cert\") pod \"frr-k8s-webhook-server-64bf5d555-xsl5f\" (UID: \"703cac15-6674-4529-9a6b-75f646a84d3e\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.921079 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vc5m\" (UniqueName: \"kubernetes.io/projected/56173127-d55d-4c05-ac7b-1cdc4484f3bb-kube-api-access-2vc5m\") pod \"frr-k8s-7dg9g\" (UID: \"56173127-d55d-4c05-ac7b-1cdc4484f3bb\") " pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.956280 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dfgp\" (UniqueName: \"kubernetes.io/projected/703cac15-6674-4529-9a6b-75f646a84d3e-kube-api-access-2dfgp\") pod \"frr-k8s-webhook-server-64bf5d555-xsl5f\" (UID: \"703cac15-6674-4529-9a6b-75f646a84d3e\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.998278 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0d474bc7-3b94-40d8-91d6-db976b2975d7-memberlist\") pod \"speaker-5qx66\" (UID: \"0d474bc7-3b94-40d8-91d6-db976b2975d7\") " pod="metallb-system/speaker-5qx66" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.998366 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f04d600-8bb7-4420-aa5f-c1b51a0a4e16-metrics-certs\") pod \"controller-68d546b9d8-vhk6s\" (UID: \"5f04d600-8bb7-4420-aa5f-c1b51a0a4e16\") " pod="metallb-system/controller-68d546b9d8-vhk6s" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.998391 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d4w7\" (UniqueName: \"kubernetes.io/projected/5f04d600-8bb7-4420-aa5f-c1b51a0a4e16-kube-api-access-6d4w7\") pod \"controller-68d546b9d8-vhk6s\" (UID: \"5f04d600-8bb7-4420-aa5f-c1b51a0a4e16\") " pod="metallb-system/controller-68d546b9d8-vhk6s" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.998409 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbgxb\" (UniqueName: \"kubernetes.io/projected/0d474bc7-3b94-40d8-91d6-db976b2975d7-kube-api-access-lbgxb\") pod \"speaker-5qx66\" (UID: \"0d474bc7-3b94-40d8-91d6-db976b2975d7\") " pod="metallb-system/speaker-5qx66" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.998434 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5f04d600-8bb7-4420-aa5f-c1b51a0a4e16-cert\") pod \"controller-68d546b9d8-vhk6s\" (UID: \"5f04d600-8bb7-4420-aa5f-c1b51a0a4e16\") " pod="metallb-system/controller-68d546b9d8-vhk6s" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.998461 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/0d474bc7-3b94-40d8-91d6-db976b2975d7-metallb-excludel2\") pod \"speaker-5qx66\" (UID: \"0d474bc7-3b94-40d8-91d6-db976b2975d7\") " pod="metallb-system/speaker-5qx66" Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.998487 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0d474bc7-3b94-40d8-91d6-db976b2975d7-metrics-certs\") pod \"speaker-5qx66\" (UID: \"0d474bc7-3b94-40d8-91d6-db976b2975d7\") " pod="metallb-system/speaker-5qx66" Oct 03 16:39:16 crc kubenswrapper[4744]: E1003 16:39:16.998839 4744 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 03 16:39:16 crc kubenswrapper[4744]: E1003 16:39:16.998925 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0d474bc7-3b94-40d8-91d6-db976b2975d7-memberlist podName:0d474bc7-3b94-40d8-91d6-db976b2975d7 nodeName:}" failed. No retries permitted until 2025-10-03 16:39:17.498902967 +0000 UTC m=+883.778778953 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/0d474bc7-3b94-40d8-91d6-db976b2975d7-memberlist") pod "speaker-5qx66" (UID: "0d474bc7-3b94-40d8-91d6-db976b2975d7") : secret "metallb-memberlist" not found Oct 03 16:39:16 crc kubenswrapper[4744]: I1003 16:39:16.999364 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/0d474bc7-3b94-40d8-91d6-db976b2975d7-metallb-excludel2\") pod \"speaker-5qx66\" (UID: \"0d474bc7-3b94-40d8-91d6-db976b2975d7\") " pod="metallb-system/speaker-5qx66" Oct 03 16:39:17 crc kubenswrapper[4744]: I1003 16:39:17.003675 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5f04d600-8bb7-4420-aa5f-c1b51a0a4e16-cert\") pod \"controller-68d546b9d8-vhk6s\" (UID: \"5f04d600-8bb7-4420-aa5f-c1b51a0a4e16\") " pod="metallb-system/controller-68d546b9d8-vhk6s" Oct 03 16:39:17 crc kubenswrapper[4744]: I1003 16:39:17.017276 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f" Oct 03 16:39:17 crc kubenswrapper[4744]: I1003 16:39:17.018073 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0d474bc7-3b94-40d8-91d6-db976b2975d7-metrics-certs\") pod \"speaker-5qx66\" (UID: \"0d474bc7-3b94-40d8-91d6-db976b2975d7\") " pod="metallb-system/speaker-5qx66" Oct 03 16:39:17 crc kubenswrapper[4744]: I1003 16:39:17.032279 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d4w7\" (UniqueName: \"kubernetes.io/projected/5f04d600-8bb7-4420-aa5f-c1b51a0a4e16-kube-api-access-6d4w7\") pod \"controller-68d546b9d8-vhk6s\" (UID: \"5f04d600-8bb7-4420-aa5f-c1b51a0a4e16\") " pod="metallb-system/controller-68d546b9d8-vhk6s" Oct 03 16:39:17 crc kubenswrapper[4744]: I1003 16:39:17.032724 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:17 crc kubenswrapper[4744]: I1003 16:39:17.034247 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbgxb\" (UniqueName: \"kubernetes.io/projected/0d474bc7-3b94-40d8-91d6-db976b2975d7-kube-api-access-lbgxb\") pod \"speaker-5qx66\" (UID: \"0d474bc7-3b94-40d8-91d6-db976b2975d7\") " pod="metallb-system/speaker-5qx66" Oct 03 16:39:17 crc kubenswrapper[4744]: I1003 16:39:17.253783 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7dg9g" event={"ID":"56173127-d55d-4c05-ac7b-1cdc4484f3bb","Type":"ContainerStarted","Data":"1ab72c8311cd412a8b50d5b71c25c78bd81876504a4cf8fc741c84c8f569c087"} Oct 03 16:39:17 crc kubenswrapper[4744]: I1003 16:39:17.462741 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f"] Oct 03 16:39:17 crc kubenswrapper[4744]: W1003 16:39:17.470037 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod703cac15_6674_4529_9a6b_75f646a84d3e.slice/crio-4d5291991aa0d7b88a78c0187ca0657e6b478e2cd5fc45cf3b8b55765d6eb0e1 WatchSource:0}: Error finding container 4d5291991aa0d7b88a78c0187ca0657e6b478e2cd5fc45cf3b8b55765d6eb0e1: Status 404 returned error can't find the container with id 4d5291991aa0d7b88a78c0187ca0657e6b478e2cd5fc45cf3b8b55765d6eb0e1 Oct 03 16:39:17 crc kubenswrapper[4744]: I1003 16:39:17.507749 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0d474bc7-3b94-40d8-91d6-db976b2975d7-memberlist\") pod \"speaker-5qx66\" (UID: \"0d474bc7-3b94-40d8-91d6-db976b2975d7\") " pod="metallb-system/speaker-5qx66" Oct 03 16:39:17 crc kubenswrapper[4744]: E1003 16:39:17.507955 4744 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 03 16:39:17 crc kubenswrapper[4744]: E1003 16:39:17.508047 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0d474bc7-3b94-40d8-91d6-db976b2975d7-memberlist podName:0d474bc7-3b94-40d8-91d6-db976b2975d7 nodeName:}" failed. No retries permitted until 2025-10-03 16:39:18.508025446 +0000 UTC m=+884.787901342 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/0d474bc7-3b94-40d8-91d6-db976b2975d7-memberlist") pod "speaker-5qx66" (UID: "0d474bc7-3b94-40d8-91d6-db976b2975d7") : secret "metallb-memberlist" not found Oct 03 16:39:17 crc kubenswrapper[4744]: I1003 16:39:17.726531 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 03 16:39:17 crc kubenswrapper[4744]: I1003 16:39:17.737935 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5f04d600-8bb7-4420-aa5f-c1b51a0a4e16-metrics-certs\") pod \"controller-68d546b9d8-vhk6s\" (UID: \"5f04d600-8bb7-4420-aa5f-c1b51a0a4e16\") " pod="metallb-system/controller-68d546b9d8-vhk6s" Oct 03 16:39:18 crc kubenswrapper[4744]: I1003 16:39:18.008339 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-vhk6s" Oct 03 16:39:18 crc kubenswrapper[4744]: I1003 16:39:18.259395 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f" event={"ID":"703cac15-6674-4529-9a6b-75f646a84d3e","Type":"ContainerStarted","Data":"4d5291991aa0d7b88a78c0187ca0657e6b478e2cd5fc45cf3b8b55765d6eb0e1"} Oct 03 16:39:18 crc kubenswrapper[4744]: I1003 16:39:18.452175 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-vhk6s"] Oct 03 16:39:18 crc kubenswrapper[4744]: I1003 16:39:18.523971 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0d474bc7-3b94-40d8-91d6-db976b2975d7-memberlist\") pod \"speaker-5qx66\" (UID: \"0d474bc7-3b94-40d8-91d6-db976b2975d7\") " pod="metallb-system/speaker-5qx66" Oct 03 16:39:18 crc kubenswrapper[4744]: E1003 16:39:18.524138 4744 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 03 16:39:18 crc kubenswrapper[4744]: E1003 16:39:18.524196 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0d474bc7-3b94-40d8-91d6-db976b2975d7-memberlist podName:0d474bc7-3b94-40d8-91d6-db976b2975d7 nodeName:}" failed. No retries permitted until 2025-10-03 16:39:20.524180892 +0000 UTC m=+886.804056778 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/0d474bc7-3b94-40d8-91d6-db976b2975d7-memberlist") pod "speaker-5qx66" (UID: "0d474bc7-3b94-40d8-91d6-db976b2975d7") : secret "metallb-memberlist" not found Oct 03 16:39:19 crc kubenswrapper[4744]: I1003 16:39:19.269442 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-vhk6s" event={"ID":"5f04d600-8bb7-4420-aa5f-c1b51a0a4e16","Type":"ContainerStarted","Data":"a10e08809e129110889d657f02809e1b60384be0267cf833da249859603be9ec"} Oct 03 16:39:19 crc kubenswrapper[4744]: I1003 16:39:19.269510 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-vhk6s" event={"ID":"5f04d600-8bb7-4420-aa5f-c1b51a0a4e16","Type":"ContainerStarted","Data":"3728a8daf1ddcd61b1ffbed509093f99cb6baf4484c5f9d4de355bd133fc26b7"} Oct 03 16:39:19 crc kubenswrapper[4744]: I1003 16:39:19.269525 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-vhk6s" event={"ID":"5f04d600-8bb7-4420-aa5f-c1b51a0a4e16","Type":"ContainerStarted","Data":"dfbccc161b7c0194e7b4745f233236c27842ec4de7b7468b2b3c2179f4d13243"} Oct 03 16:39:19 crc kubenswrapper[4744]: I1003 16:39:19.269724 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-vhk6s" Oct 03 16:39:19 crc kubenswrapper[4744]: I1003 16:39:19.297277 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-vhk6s" podStartSLOduration=3.297253817 podStartE2EDuration="3.297253817s" podCreationTimestamp="2025-10-03 16:39:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:39:19.292633908 +0000 UTC m=+885.572509804" watchObservedRunningTime="2025-10-03 16:39:19.297253817 +0000 UTC m=+885.577129713" Oct 03 16:39:20 crc kubenswrapper[4744]: I1003 16:39:20.554358 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0d474bc7-3b94-40d8-91d6-db976b2975d7-memberlist\") pod \"speaker-5qx66\" (UID: \"0d474bc7-3b94-40d8-91d6-db976b2975d7\") " pod="metallb-system/speaker-5qx66" Oct 03 16:39:20 crc kubenswrapper[4744]: I1003 16:39:20.579519 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0d474bc7-3b94-40d8-91d6-db976b2975d7-memberlist\") pod \"speaker-5qx66\" (UID: \"0d474bc7-3b94-40d8-91d6-db976b2975d7\") " pod="metallb-system/speaker-5qx66" Oct 03 16:39:20 crc kubenswrapper[4744]: I1003 16:39:20.737307 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-5qx66" Oct 03 16:39:20 crc kubenswrapper[4744]: W1003 16:39:20.801009 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d474bc7_3b94_40d8_91d6_db976b2975d7.slice/crio-565504744bc8efd9c03c09d819ddfe4ac9b6785a730586964215feab5bdb699c WatchSource:0}: Error finding container 565504744bc8efd9c03c09d819ddfe4ac9b6785a730586964215feab5bdb699c: Status 404 returned error can't find the container with id 565504744bc8efd9c03c09d819ddfe4ac9b6785a730586964215feab5bdb699c Oct 03 16:39:21 crc kubenswrapper[4744]: I1003 16:39:21.302840 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5qx66" event={"ID":"0d474bc7-3b94-40d8-91d6-db976b2975d7","Type":"ContainerStarted","Data":"dc4bfef7a21fae483bca3052e65b400346372a1ec71f4bd672a64a03c58c63f3"} Oct 03 16:39:21 crc kubenswrapper[4744]: I1003 16:39:21.303231 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5qx66" event={"ID":"0d474bc7-3b94-40d8-91d6-db976b2975d7","Type":"ContainerStarted","Data":"565504744bc8efd9c03c09d819ddfe4ac9b6785a730586964215feab5bdb699c"} Oct 03 16:39:22 crc kubenswrapper[4744]: I1003 16:39:22.311276 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-5qx66" event={"ID":"0d474bc7-3b94-40d8-91d6-db976b2975d7","Type":"ContainerStarted","Data":"a4b0dac6545d0b3f50f1d92b934c93c17ba622212bdb0c2ef33800f440c5bc80"} Oct 03 16:39:22 crc kubenswrapper[4744]: I1003 16:39:22.311423 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-5qx66" Oct 03 16:39:22 crc kubenswrapper[4744]: I1003 16:39:22.329695 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-5qx66" podStartSLOduration=6.32967555 podStartE2EDuration="6.32967555s" podCreationTimestamp="2025-10-03 16:39:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:39:22.327828922 +0000 UTC m=+888.607704848" watchObservedRunningTime="2025-10-03 16:39:22.32967555 +0000 UTC m=+888.609551446" Oct 03 16:39:25 crc kubenswrapper[4744]: I1003 16:39:25.334617 4744 generic.go:334] "Generic (PLEG): container finished" podID="56173127-d55d-4c05-ac7b-1cdc4484f3bb" containerID="b936330095deb50cda48fb770b2e189e77781557def54cb5438fd1d264f2044c" exitCode=0 Oct 03 16:39:25 crc kubenswrapper[4744]: I1003 16:39:25.334704 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7dg9g" event={"ID":"56173127-d55d-4c05-ac7b-1cdc4484f3bb","Type":"ContainerDied","Data":"b936330095deb50cda48fb770b2e189e77781557def54cb5438fd1d264f2044c"} Oct 03 16:39:25 crc kubenswrapper[4744]: I1003 16:39:25.336670 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f" event={"ID":"703cac15-6674-4529-9a6b-75f646a84d3e","Type":"ContainerStarted","Data":"1866babf9aaf76b5ac3695e2a68ad27a8390583d2d39c66f177f15f50700bb2e"} Oct 03 16:39:25 crc kubenswrapper[4744]: I1003 16:39:25.336859 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f" Oct 03 16:39:25 crc kubenswrapper[4744]: I1003 16:39:25.379067 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f" podStartSLOduration=1.799487068 podStartE2EDuration="9.379047058s" podCreationTimestamp="2025-10-03 16:39:16 +0000 UTC" firstStartedPulling="2025-10-03 16:39:17.473083389 +0000 UTC m=+883.752959295" lastFinishedPulling="2025-10-03 16:39:25.052643399 +0000 UTC m=+891.332519285" observedRunningTime="2025-10-03 16:39:25.375655951 +0000 UTC m=+891.655531877" watchObservedRunningTime="2025-10-03 16:39:25.379047058 +0000 UTC m=+891.658922954" Oct 03 16:39:26 crc kubenswrapper[4744]: I1003 16:39:26.347698 4744 generic.go:334] "Generic (PLEG): container finished" podID="56173127-d55d-4c05-ac7b-1cdc4484f3bb" containerID="200c2e36ce94c5c583ddfbc65188be211096f14a2ef76e6f81802a74f5770c32" exitCode=0 Oct 03 16:39:26 crc kubenswrapper[4744]: I1003 16:39:26.347897 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7dg9g" event={"ID":"56173127-d55d-4c05-ac7b-1cdc4484f3bb","Type":"ContainerDied","Data":"200c2e36ce94c5c583ddfbc65188be211096f14a2ef76e6f81802a74f5770c32"} Oct 03 16:39:27 crc kubenswrapper[4744]: I1003 16:39:27.356294 4744 generic.go:334] "Generic (PLEG): container finished" podID="56173127-d55d-4c05-ac7b-1cdc4484f3bb" containerID="6f52a2eb7945cf316254c86c5aff60f55492c8f0a1a0cc28c0eb08e64b7bfedb" exitCode=0 Oct 03 16:39:27 crc kubenswrapper[4744]: I1003 16:39:27.356393 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7dg9g" event={"ID":"56173127-d55d-4c05-ac7b-1cdc4484f3bb","Type":"ContainerDied","Data":"6f52a2eb7945cf316254c86c5aff60f55492c8f0a1a0cc28c0eb08e64b7bfedb"} Oct 03 16:39:28 crc kubenswrapper[4744]: I1003 16:39:28.016015 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-vhk6s" Oct 03 16:39:28 crc kubenswrapper[4744]: I1003 16:39:28.367993 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7dg9g" event={"ID":"56173127-d55d-4c05-ac7b-1cdc4484f3bb","Type":"ContainerStarted","Data":"605bbc89643d9bfe1e390c87d01a0fa71c9fd5fe1933bc14cac7d15de244b0f1"} Oct 03 16:39:28 crc kubenswrapper[4744]: I1003 16:39:28.368053 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7dg9g" event={"ID":"56173127-d55d-4c05-ac7b-1cdc4484f3bb","Type":"ContainerStarted","Data":"b04efcf2027304cf853611160449c1970c08192e621302d3bc8cc76a299bd9bc"} Oct 03 16:39:28 crc kubenswrapper[4744]: I1003 16:39:28.368068 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7dg9g" event={"ID":"56173127-d55d-4c05-ac7b-1cdc4484f3bb","Type":"ContainerStarted","Data":"7cf4e46caf105f618081c59bbc4eec58bd99ad2f62820a3dcb753bd3be5ec8da"} Oct 03 16:39:28 crc kubenswrapper[4744]: I1003 16:39:28.368078 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7dg9g" event={"ID":"56173127-d55d-4c05-ac7b-1cdc4484f3bb","Type":"ContainerStarted","Data":"b9dc2783760ec7987fa389df6a73731e976d7f5e7faacc63dda6f0ede19684e0"} Oct 03 16:39:28 crc kubenswrapper[4744]: I1003 16:39:28.368090 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7dg9g" event={"ID":"56173127-d55d-4c05-ac7b-1cdc4484f3bb","Type":"ContainerStarted","Data":"65579dd8eb10132731ffe9d4dedd7aa709fdf48bf1df8fa239066d06b76794d9"} Oct 03 16:39:28 crc kubenswrapper[4744]: I1003 16:39:28.368102 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7dg9g" event={"ID":"56173127-d55d-4c05-ac7b-1cdc4484f3bb","Type":"ContainerStarted","Data":"46c51c79ca9ff7ecf8d028660e971b136a981dffc3e10e8625bb252c91456f67"} Oct 03 16:39:28 crc kubenswrapper[4744]: I1003 16:39:28.368373 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:28 crc kubenswrapper[4744]: I1003 16:39:28.405475 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-7dg9g" podStartSLOduration=4.555359073 podStartE2EDuration="12.405445917s" podCreationTimestamp="2025-10-03 16:39:16 +0000 UTC" firstStartedPulling="2025-10-03 16:39:17.163677107 +0000 UTC m=+883.443553003" lastFinishedPulling="2025-10-03 16:39:25.013763951 +0000 UTC m=+891.293639847" observedRunningTime="2025-10-03 16:39:28.402595823 +0000 UTC m=+894.682471759" watchObservedRunningTime="2025-10-03 16:39:28.405445917 +0000 UTC m=+894.685321843" Oct 03 16:39:32 crc kubenswrapper[4744]: I1003 16:39:32.033428 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:32 crc kubenswrapper[4744]: I1003 16:39:32.072924 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:37 crc kubenswrapper[4744]: I1003 16:39:37.024598 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xsl5f" Oct 03 16:39:37 crc kubenswrapper[4744]: I1003 16:39:37.037911 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-7dg9g" Oct 03 16:39:40 crc kubenswrapper[4744]: I1003 16:39:40.742566 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-5qx66" Oct 03 16:39:43 crc kubenswrapper[4744]: I1003 16:39:43.564256 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-hd9xm"] Oct 03 16:39:43 crc kubenswrapper[4744]: I1003 16:39:43.565211 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hd9xm" Oct 03 16:39:43 crc kubenswrapper[4744]: I1003 16:39:43.569334 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 03 16:39:43 crc kubenswrapper[4744]: I1003 16:39:43.569381 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 03 16:39:43 crc kubenswrapper[4744]: I1003 16:39:43.569797 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-xh7n9" Oct 03 16:39:43 crc kubenswrapper[4744]: I1003 16:39:43.587728 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hd9xm"] Oct 03 16:39:43 crc kubenswrapper[4744]: I1003 16:39:43.624479 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llj66\" (UniqueName: \"kubernetes.io/projected/0c64c85d-2239-4946-b6c4-88d697ffd6e1-kube-api-access-llj66\") pod \"openstack-operator-index-hd9xm\" (UID: \"0c64c85d-2239-4946-b6c4-88d697ffd6e1\") " pod="openstack-operators/openstack-operator-index-hd9xm" Oct 03 16:39:43 crc kubenswrapper[4744]: I1003 16:39:43.726368 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llj66\" (UniqueName: \"kubernetes.io/projected/0c64c85d-2239-4946-b6c4-88d697ffd6e1-kube-api-access-llj66\") pod \"openstack-operator-index-hd9xm\" (UID: \"0c64c85d-2239-4946-b6c4-88d697ffd6e1\") " pod="openstack-operators/openstack-operator-index-hd9xm" Oct 03 16:39:43 crc kubenswrapper[4744]: I1003 16:39:43.748758 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llj66\" (UniqueName: \"kubernetes.io/projected/0c64c85d-2239-4946-b6c4-88d697ffd6e1-kube-api-access-llj66\") pod \"openstack-operator-index-hd9xm\" (UID: \"0c64c85d-2239-4946-b6c4-88d697ffd6e1\") " pod="openstack-operators/openstack-operator-index-hd9xm" Oct 03 16:39:43 crc kubenswrapper[4744]: I1003 16:39:43.895478 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hd9xm" Oct 03 16:39:44 crc kubenswrapper[4744]: I1003 16:39:44.148898 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hd9xm"] Oct 03 16:39:44 crc kubenswrapper[4744]: W1003 16:39:44.154781 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c64c85d_2239_4946_b6c4_88d697ffd6e1.slice/crio-100f971181a0723f209db3d65322007d1421d127ccdbcf1d52c9f04e71e8d8e6 WatchSource:0}: Error finding container 100f971181a0723f209db3d65322007d1421d127ccdbcf1d52c9f04e71e8d8e6: Status 404 returned error can't find the container with id 100f971181a0723f209db3d65322007d1421d127ccdbcf1d52c9f04e71e8d8e6 Oct 03 16:39:44 crc kubenswrapper[4744]: I1003 16:39:44.490848 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hd9xm" event={"ID":"0c64c85d-2239-4946-b6c4-88d697ffd6e1","Type":"ContainerStarted","Data":"100f971181a0723f209db3d65322007d1421d127ccdbcf1d52c9f04e71e8d8e6"} Oct 03 16:39:46 crc kubenswrapper[4744]: I1003 16:39:46.509932 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hd9xm" event={"ID":"0c64c85d-2239-4946-b6c4-88d697ffd6e1","Type":"ContainerStarted","Data":"2201ec7106903feec5b39ffcdb0930582320aa6768d975475cff601755647f0e"} Oct 03 16:39:46 crc kubenswrapper[4744]: I1003 16:39:46.528130 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-hd9xm" podStartSLOduration=1.574918776 podStartE2EDuration="3.528093094s" podCreationTimestamp="2025-10-03 16:39:43 +0000 UTC" firstStartedPulling="2025-10-03 16:39:44.156933206 +0000 UTC m=+910.436809112" lastFinishedPulling="2025-10-03 16:39:46.110107534 +0000 UTC m=+912.389983430" observedRunningTime="2025-10-03 16:39:46.52558019 +0000 UTC m=+912.805456166" watchObservedRunningTime="2025-10-03 16:39:46.528093094 +0000 UTC m=+912.807969030" Oct 03 16:39:46 crc kubenswrapper[4744]: I1003 16:39:46.935665 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-hd9xm"] Oct 03 16:39:47 crc kubenswrapper[4744]: I1003 16:39:47.544368 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-n2bfv"] Oct 03 16:39:47 crc kubenswrapper[4744]: I1003 16:39:47.545613 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n2bfv" Oct 03 16:39:47 crc kubenswrapper[4744]: I1003 16:39:47.550517 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-n2bfv"] Oct 03 16:39:47 crc kubenswrapper[4744]: I1003 16:39:47.588387 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcmpx\" (UniqueName: \"kubernetes.io/projected/d8b895bc-8e19-4b92-999f-d579c1d51e51-kube-api-access-vcmpx\") pod \"openstack-operator-index-n2bfv\" (UID: \"d8b895bc-8e19-4b92-999f-d579c1d51e51\") " pod="openstack-operators/openstack-operator-index-n2bfv" Oct 03 16:39:47 crc kubenswrapper[4744]: I1003 16:39:47.689597 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcmpx\" (UniqueName: \"kubernetes.io/projected/d8b895bc-8e19-4b92-999f-d579c1d51e51-kube-api-access-vcmpx\") pod \"openstack-operator-index-n2bfv\" (UID: \"d8b895bc-8e19-4b92-999f-d579c1d51e51\") " pod="openstack-operators/openstack-operator-index-n2bfv" Oct 03 16:39:47 crc kubenswrapper[4744]: I1003 16:39:47.709022 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcmpx\" (UniqueName: \"kubernetes.io/projected/d8b895bc-8e19-4b92-999f-d579c1d51e51-kube-api-access-vcmpx\") pod \"openstack-operator-index-n2bfv\" (UID: \"d8b895bc-8e19-4b92-999f-d579c1d51e51\") " pod="openstack-operators/openstack-operator-index-n2bfv" Oct 03 16:39:47 crc kubenswrapper[4744]: I1003 16:39:47.875615 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n2bfv" Oct 03 16:39:48 crc kubenswrapper[4744]: I1003 16:39:48.162894 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-n2bfv"] Oct 03 16:39:48 crc kubenswrapper[4744]: W1003 16:39:48.170985 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8b895bc_8e19_4b92_999f_d579c1d51e51.slice/crio-3afcc7a3e52c2a50db6e6af247ca52f8a65e855e5ab4581afc092130cd6d5efe WatchSource:0}: Error finding container 3afcc7a3e52c2a50db6e6af247ca52f8a65e855e5ab4581afc092130cd6d5efe: Status 404 returned error can't find the container with id 3afcc7a3e52c2a50db6e6af247ca52f8a65e855e5ab4581afc092130cd6d5efe Oct 03 16:39:48 crc kubenswrapper[4744]: I1003 16:39:48.525158 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n2bfv" event={"ID":"d8b895bc-8e19-4b92-999f-d579c1d51e51","Type":"ContainerStarted","Data":"01842e2c87e773b275baf7ce0ec840a4146fd9a96995f3ab7cf233e3754ae0fc"} Oct 03 16:39:48 crc kubenswrapper[4744]: I1003 16:39:48.525201 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-hd9xm" podUID="0c64c85d-2239-4946-b6c4-88d697ffd6e1" containerName="registry-server" containerID="cri-o://2201ec7106903feec5b39ffcdb0930582320aa6768d975475cff601755647f0e" gracePeriod=2 Oct 03 16:39:48 crc kubenswrapper[4744]: I1003 16:39:48.525840 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n2bfv" event={"ID":"d8b895bc-8e19-4b92-999f-d579c1d51e51","Type":"ContainerStarted","Data":"3afcc7a3e52c2a50db6e6af247ca52f8a65e855e5ab4581afc092130cd6d5efe"} Oct 03 16:39:48 crc kubenswrapper[4744]: I1003 16:39:48.551862 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-n2bfv" podStartSLOduration=1.482479344 podStartE2EDuration="1.551845404s" podCreationTimestamp="2025-10-03 16:39:47 +0000 UTC" firstStartedPulling="2025-10-03 16:39:48.175169905 +0000 UTC m=+914.455045811" lastFinishedPulling="2025-10-03 16:39:48.244535975 +0000 UTC m=+914.524411871" observedRunningTime="2025-10-03 16:39:48.550589842 +0000 UTC m=+914.830465738" watchObservedRunningTime="2025-10-03 16:39:48.551845404 +0000 UTC m=+914.831721300" Oct 03 16:39:48 crc kubenswrapper[4744]: I1003 16:39:48.973563 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hd9xm" Oct 03 16:39:49 crc kubenswrapper[4744]: I1003 16:39:49.014846 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llj66\" (UniqueName: \"kubernetes.io/projected/0c64c85d-2239-4946-b6c4-88d697ffd6e1-kube-api-access-llj66\") pod \"0c64c85d-2239-4946-b6c4-88d697ffd6e1\" (UID: \"0c64c85d-2239-4946-b6c4-88d697ffd6e1\") " Oct 03 16:39:49 crc kubenswrapper[4744]: I1003 16:39:49.027433 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c64c85d-2239-4946-b6c4-88d697ffd6e1-kube-api-access-llj66" (OuterVolumeSpecName: "kube-api-access-llj66") pod "0c64c85d-2239-4946-b6c4-88d697ffd6e1" (UID: "0c64c85d-2239-4946-b6c4-88d697ffd6e1"). InnerVolumeSpecName "kube-api-access-llj66". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:39:49 crc kubenswrapper[4744]: I1003 16:39:49.116009 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llj66\" (UniqueName: \"kubernetes.io/projected/0c64c85d-2239-4946-b6c4-88d697ffd6e1-kube-api-access-llj66\") on node \"crc\" DevicePath \"\"" Oct 03 16:39:49 crc kubenswrapper[4744]: I1003 16:39:49.533939 4744 generic.go:334] "Generic (PLEG): container finished" podID="0c64c85d-2239-4946-b6c4-88d697ffd6e1" containerID="2201ec7106903feec5b39ffcdb0930582320aa6768d975475cff601755647f0e" exitCode=0 Oct 03 16:39:49 crc kubenswrapper[4744]: I1003 16:39:49.534019 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hd9xm" Oct 03 16:39:49 crc kubenswrapper[4744]: I1003 16:39:49.533999 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hd9xm" event={"ID":"0c64c85d-2239-4946-b6c4-88d697ffd6e1","Type":"ContainerDied","Data":"2201ec7106903feec5b39ffcdb0930582320aa6768d975475cff601755647f0e"} Oct 03 16:39:49 crc kubenswrapper[4744]: I1003 16:39:49.534092 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hd9xm" event={"ID":"0c64c85d-2239-4946-b6c4-88d697ffd6e1","Type":"ContainerDied","Data":"100f971181a0723f209db3d65322007d1421d127ccdbcf1d52c9f04e71e8d8e6"} Oct 03 16:39:49 crc kubenswrapper[4744]: I1003 16:39:49.534133 4744 scope.go:117] "RemoveContainer" containerID="2201ec7106903feec5b39ffcdb0930582320aa6768d975475cff601755647f0e" Oct 03 16:39:49 crc kubenswrapper[4744]: I1003 16:39:49.559783 4744 scope.go:117] "RemoveContainer" containerID="2201ec7106903feec5b39ffcdb0930582320aa6768d975475cff601755647f0e" Oct 03 16:39:49 crc kubenswrapper[4744]: E1003 16:39:49.560206 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2201ec7106903feec5b39ffcdb0930582320aa6768d975475cff601755647f0e\": container with ID starting with 2201ec7106903feec5b39ffcdb0930582320aa6768d975475cff601755647f0e not found: ID does not exist" containerID="2201ec7106903feec5b39ffcdb0930582320aa6768d975475cff601755647f0e" Oct 03 16:39:49 crc kubenswrapper[4744]: I1003 16:39:49.560251 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2201ec7106903feec5b39ffcdb0930582320aa6768d975475cff601755647f0e"} err="failed to get container status \"2201ec7106903feec5b39ffcdb0930582320aa6768d975475cff601755647f0e\": rpc error: code = NotFound desc = could not find container \"2201ec7106903feec5b39ffcdb0930582320aa6768d975475cff601755647f0e\": container with ID starting with 2201ec7106903feec5b39ffcdb0930582320aa6768d975475cff601755647f0e not found: ID does not exist" Oct 03 16:39:49 crc kubenswrapper[4744]: I1003 16:39:49.578999 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-hd9xm"] Oct 03 16:39:49 crc kubenswrapper[4744]: I1003 16:39:49.582880 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-hd9xm"] Oct 03 16:39:50 crc kubenswrapper[4744]: I1003 16:39:50.899094 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c64c85d-2239-4946-b6c4-88d697ffd6e1" path="/var/lib/kubelet/pods/0c64c85d-2239-4946-b6c4-88d697ffd6e1/volumes" Oct 03 16:39:57 crc kubenswrapper[4744]: I1003 16:39:57.876523 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-n2bfv" Oct 03 16:39:57 crc kubenswrapper[4744]: I1003 16:39:57.876892 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-n2bfv" Oct 03 16:39:57 crc kubenswrapper[4744]: I1003 16:39:57.903378 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-n2bfv" Oct 03 16:39:58 crc kubenswrapper[4744]: I1003 16:39:58.643341 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-n2bfv" Oct 03 16:40:03 crc kubenswrapper[4744]: I1003 16:40:03.773481 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx"] Oct 03 16:40:03 crc kubenswrapper[4744]: E1003 16:40:03.774024 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c64c85d-2239-4946-b6c4-88d697ffd6e1" containerName="registry-server" Oct 03 16:40:03 crc kubenswrapper[4744]: I1003 16:40:03.774036 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c64c85d-2239-4946-b6c4-88d697ffd6e1" containerName="registry-server" Oct 03 16:40:03 crc kubenswrapper[4744]: I1003 16:40:03.774150 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c64c85d-2239-4946-b6c4-88d697ffd6e1" containerName="registry-server" Oct 03 16:40:03 crc kubenswrapper[4744]: I1003 16:40:03.774973 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" Oct 03 16:40:03 crc kubenswrapper[4744]: I1003 16:40:03.778901 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-zq272" Oct 03 16:40:03 crc kubenswrapper[4744]: I1003 16:40:03.796132 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx"] Oct 03 16:40:03 crc kubenswrapper[4744]: I1003 16:40:03.833201 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aaa2ef65-fb80-4f95-81fc-29841e8b6847-util\") pod \"43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx\" (UID: \"aaa2ef65-fb80-4f95-81fc-29841e8b6847\") " pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" Oct 03 16:40:03 crc kubenswrapper[4744]: I1003 16:40:03.833277 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t62sk\" (UniqueName: \"kubernetes.io/projected/aaa2ef65-fb80-4f95-81fc-29841e8b6847-kube-api-access-t62sk\") pod \"43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx\" (UID: \"aaa2ef65-fb80-4f95-81fc-29841e8b6847\") " pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" Oct 03 16:40:03 crc kubenswrapper[4744]: I1003 16:40:03.833327 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aaa2ef65-fb80-4f95-81fc-29841e8b6847-bundle\") pod \"43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx\" (UID: \"aaa2ef65-fb80-4f95-81fc-29841e8b6847\") " pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" Oct 03 16:40:03 crc kubenswrapper[4744]: I1003 16:40:03.934987 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aaa2ef65-fb80-4f95-81fc-29841e8b6847-util\") pod \"43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx\" (UID: \"aaa2ef65-fb80-4f95-81fc-29841e8b6847\") " pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" Oct 03 16:40:03 crc kubenswrapper[4744]: I1003 16:40:03.935096 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t62sk\" (UniqueName: \"kubernetes.io/projected/aaa2ef65-fb80-4f95-81fc-29841e8b6847-kube-api-access-t62sk\") pod \"43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx\" (UID: \"aaa2ef65-fb80-4f95-81fc-29841e8b6847\") " pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" Oct 03 16:40:03 crc kubenswrapper[4744]: I1003 16:40:03.935161 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aaa2ef65-fb80-4f95-81fc-29841e8b6847-bundle\") pod \"43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx\" (UID: \"aaa2ef65-fb80-4f95-81fc-29841e8b6847\") " pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" Oct 03 16:40:03 crc kubenswrapper[4744]: I1003 16:40:03.935982 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aaa2ef65-fb80-4f95-81fc-29841e8b6847-bundle\") pod \"43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx\" (UID: \"aaa2ef65-fb80-4f95-81fc-29841e8b6847\") " pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" Oct 03 16:40:03 crc kubenswrapper[4744]: I1003 16:40:03.935998 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aaa2ef65-fb80-4f95-81fc-29841e8b6847-util\") pod \"43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx\" (UID: \"aaa2ef65-fb80-4f95-81fc-29841e8b6847\") " pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" Oct 03 16:40:03 crc kubenswrapper[4744]: I1003 16:40:03.957710 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t62sk\" (UniqueName: \"kubernetes.io/projected/aaa2ef65-fb80-4f95-81fc-29841e8b6847-kube-api-access-t62sk\") pod \"43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx\" (UID: \"aaa2ef65-fb80-4f95-81fc-29841e8b6847\") " pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" Oct 03 16:40:04 crc kubenswrapper[4744]: I1003 16:40:04.097037 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" Oct 03 16:40:04 crc kubenswrapper[4744]: I1003 16:40:04.317100 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx"] Oct 03 16:40:04 crc kubenswrapper[4744]: W1003 16:40:04.322996 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaaa2ef65_fb80_4f95_81fc_29841e8b6847.slice/crio-dbe3576f5cb2a9f15cbaa063aca4dd0aa85db8888b48d5b3c0e21d0416edf3cf WatchSource:0}: Error finding container dbe3576f5cb2a9f15cbaa063aca4dd0aa85db8888b48d5b3c0e21d0416edf3cf: Status 404 returned error can't find the container with id dbe3576f5cb2a9f15cbaa063aca4dd0aa85db8888b48d5b3c0e21d0416edf3cf Oct 03 16:40:04 crc kubenswrapper[4744]: I1003 16:40:04.643656 4744 generic.go:334] "Generic (PLEG): container finished" podID="aaa2ef65-fb80-4f95-81fc-29841e8b6847" containerID="450318202e24cb917a3d77638374615aa3901aeff3efd8acf76838fff07b6f79" exitCode=0 Oct 03 16:40:04 crc kubenswrapper[4744]: I1003 16:40:04.643731 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" event={"ID":"aaa2ef65-fb80-4f95-81fc-29841e8b6847","Type":"ContainerDied","Data":"450318202e24cb917a3d77638374615aa3901aeff3efd8acf76838fff07b6f79"} Oct 03 16:40:04 crc kubenswrapper[4744]: I1003 16:40:04.643946 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" event={"ID":"aaa2ef65-fb80-4f95-81fc-29841e8b6847","Type":"ContainerStarted","Data":"dbe3576f5cb2a9f15cbaa063aca4dd0aa85db8888b48d5b3c0e21d0416edf3cf"} Oct 03 16:40:04 crc kubenswrapper[4744]: I1003 16:40:04.646004 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 16:40:04 crc kubenswrapper[4744]: I1003 16:40:04.668952 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:40:04 crc kubenswrapper[4744]: I1003 16:40:04.669012 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:40:05 crc kubenswrapper[4744]: I1003 16:40:05.654984 4744 generic.go:334] "Generic (PLEG): container finished" podID="aaa2ef65-fb80-4f95-81fc-29841e8b6847" containerID="48648b12d189ae52f8ef76e7307120891e892184f85b16c56be619956bd4626e" exitCode=0 Oct 03 16:40:05 crc kubenswrapper[4744]: I1003 16:40:05.655384 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" event={"ID":"aaa2ef65-fb80-4f95-81fc-29841e8b6847","Type":"ContainerDied","Data":"48648b12d189ae52f8ef76e7307120891e892184f85b16c56be619956bd4626e"} Oct 03 16:40:06 crc kubenswrapper[4744]: I1003 16:40:06.663633 4744 generic.go:334] "Generic (PLEG): container finished" podID="aaa2ef65-fb80-4f95-81fc-29841e8b6847" containerID="06009cf20e1f13ae1812fd801d0cd3907e36f669d81a1f7cfb04d190ac29ffc3" exitCode=0 Oct 03 16:40:06 crc kubenswrapper[4744]: I1003 16:40:06.663679 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" event={"ID":"aaa2ef65-fb80-4f95-81fc-29841e8b6847","Type":"ContainerDied","Data":"06009cf20e1f13ae1812fd801d0cd3907e36f669d81a1f7cfb04d190ac29ffc3"} Oct 03 16:40:07 crc kubenswrapper[4744]: I1003 16:40:07.943038 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" Oct 03 16:40:08 crc kubenswrapper[4744]: I1003 16:40:08.007414 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aaa2ef65-fb80-4f95-81fc-29841e8b6847-bundle\") pod \"aaa2ef65-fb80-4f95-81fc-29841e8b6847\" (UID: \"aaa2ef65-fb80-4f95-81fc-29841e8b6847\") " Oct 03 16:40:08 crc kubenswrapper[4744]: I1003 16:40:08.007470 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t62sk\" (UniqueName: \"kubernetes.io/projected/aaa2ef65-fb80-4f95-81fc-29841e8b6847-kube-api-access-t62sk\") pod \"aaa2ef65-fb80-4f95-81fc-29841e8b6847\" (UID: \"aaa2ef65-fb80-4f95-81fc-29841e8b6847\") " Oct 03 16:40:08 crc kubenswrapper[4744]: I1003 16:40:08.007521 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aaa2ef65-fb80-4f95-81fc-29841e8b6847-util\") pod \"aaa2ef65-fb80-4f95-81fc-29841e8b6847\" (UID: \"aaa2ef65-fb80-4f95-81fc-29841e8b6847\") " Oct 03 16:40:08 crc kubenswrapper[4744]: I1003 16:40:08.008834 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaa2ef65-fb80-4f95-81fc-29841e8b6847-bundle" (OuterVolumeSpecName: "bundle") pod "aaa2ef65-fb80-4f95-81fc-29841e8b6847" (UID: "aaa2ef65-fb80-4f95-81fc-29841e8b6847"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:40:08 crc kubenswrapper[4744]: I1003 16:40:08.018743 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaa2ef65-fb80-4f95-81fc-29841e8b6847-kube-api-access-t62sk" (OuterVolumeSpecName: "kube-api-access-t62sk") pod "aaa2ef65-fb80-4f95-81fc-29841e8b6847" (UID: "aaa2ef65-fb80-4f95-81fc-29841e8b6847"). InnerVolumeSpecName "kube-api-access-t62sk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:40:08 crc kubenswrapper[4744]: I1003 16:40:08.024138 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaa2ef65-fb80-4f95-81fc-29841e8b6847-util" (OuterVolumeSpecName: "util") pod "aaa2ef65-fb80-4f95-81fc-29841e8b6847" (UID: "aaa2ef65-fb80-4f95-81fc-29841e8b6847"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:40:08 crc kubenswrapper[4744]: I1003 16:40:08.108880 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t62sk\" (UniqueName: \"kubernetes.io/projected/aaa2ef65-fb80-4f95-81fc-29841e8b6847-kube-api-access-t62sk\") on node \"crc\" DevicePath \"\"" Oct 03 16:40:08 crc kubenswrapper[4744]: I1003 16:40:08.108915 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aaa2ef65-fb80-4f95-81fc-29841e8b6847-util\") on node \"crc\" DevicePath \"\"" Oct 03 16:40:08 crc kubenswrapper[4744]: I1003 16:40:08.108929 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aaa2ef65-fb80-4f95-81fc-29841e8b6847-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:40:08 crc kubenswrapper[4744]: I1003 16:40:08.683267 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" event={"ID":"aaa2ef65-fb80-4f95-81fc-29841e8b6847","Type":"ContainerDied","Data":"dbe3576f5cb2a9f15cbaa063aca4dd0aa85db8888b48d5b3c0e21d0416edf3cf"} Oct 03 16:40:08 crc kubenswrapper[4744]: I1003 16:40:08.683320 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx" Oct 03 16:40:08 crc kubenswrapper[4744]: I1003 16:40:08.683334 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbe3576f5cb2a9f15cbaa063aca4dd0aa85db8888b48d5b3c0e21d0416edf3cf" Oct 03 16:40:16 crc kubenswrapper[4744]: I1003 16:40:16.474013 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-d94d76459-krksv"] Oct 03 16:40:16 crc kubenswrapper[4744]: E1003 16:40:16.474625 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaa2ef65-fb80-4f95-81fc-29841e8b6847" containerName="pull" Oct 03 16:40:16 crc kubenswrapper[4744]: I1003 16:40:16.474639 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaa2ef65-fb80-4f95-81fc-29841e8b6847" containerName="pull" Oct 03 16:40:16 crc kubenswrapper[4744]: E1003 16:40:16.474652 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaa2ef65-fb80-4f95-81fc-29841e8b6847" containerName="extract" Oct 03 16:40:16 crc kubenswrapper[4744]: I1003 16:40:16.474659 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaa2ef65-fb80-4f95-81fc-29841e8b6847" containerName="extract" Oct 03 16:40:16 crc kubenswrapper[4744]: E1003 16:40:16.474666 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaa2ef65-fb80-4f95-81fc-29841e8b6847" containerName="util" Oct 03 16:40:16 crc kubenswrapper[4744]: I1003 16:40:16.474672 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaa2ef65-fb80-4f95-81fc-29841e8b6847" containerName="util" Oct 03 16:40:16 crc kubenswrapper[4744]: I1003 16:40:16.474812 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaa2ef65-fb80-4f95-81fc-29841e8b6847" containerName="extract" Oct 03 16:40:16 crc kubenswrapper[4744]: I1003 16:40:16.475448 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-d94d76459-krksv" Oct 03 16:40:16 crc kubenswrapper[4744]: I1003 16:40:16.477972 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-qg526" Oct 03 16:40:16 crc kubenswrapper[4744]: I1003 16:40:16.503799 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-d94d76459-krksv"] Oct 03 16:40:16 crc kubenswrapper[4744]: I1003 16:40:16.534663 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2mpz\" (UniqueName: \"kubernetes.io/projected/df419359-3ee4-4c00-bb37-c7387b19492a-kube-api-access-q2mpz\") pod \"openstack-operator-controller-operator-d94d76459-krksv\" (UID: \"df419359-3ee4-4c00-bb37-c7387b19492a\") " pod="openstack-operators/openstack-operator-controller-operator-d94d76459-krksv" Oct 03 16:40:16 crc kubenswrapper[4744]: I1003 16:40:16.636431 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2mpz\" (UniqueName: \"kubernetes.io/projected/df419359-3ee4-4c00-bb37-c7387b19492a-kube-api-access-q2mpz\") pod \"openstack-operator-controller-operator-d94d76459-krksv\" (UID: \"df419359-3ee4-4c00-bb37-c7387b19492a\") " pod="openstack-operators/openstack-operator-controller-operator-d94d76459-krksv" Oct 03 16:40:16 crc kubenswrapper[4744]: I1003 16:40:16.660567 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2mpz\" (UniqueName: \"kubernetes.io/projected/df419359-3ee4-4c00-bb37-c7387b19492a-kube-api-access-q2mpz\") pod \"openstack-operator-controller-operator-d94d76459-krksv\" (UID: \"df419359-3ee4-4c00-bb37-c7387b19492a\") " pod="openstack-operators/openstack-operator-controller-operator-d94d76459-krksv" Oct 03 16:40:16 crc kubenswrapper[4744]: I1003 16:40:16.792021 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-d94d76459-krksv" Oct 03 16:40:17 crc kubenswrapper[4744]: I1003 16:40:17.212840 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-d94d76459-krksv"] Oct 03 16:40:17 crc kubenswrapper[4744]: I1003 16:40:17.757916 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-d94d76459-krksv" event={"ID":"df419359-3ee4-4c00-bb37-c7387b19492a","Type":"ContainerStarted","Data":"af8ee9de7411117ec5c968e598dbf689bc46d8c81d9fa10d506d846928efe1d8"} Oct 03 16:40:21 crc kubenswrapper[4744]: I1003 16:40:21.784951 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-d94d76459-krksv" event={"ID":"df419359-3ee4-4c00-bb37-c7387b19492a","Type":"ContainerStarted","Data":"8c79eb49eaee5fd19b5799008b7645724269de3a811f3f93743085d6b137a80c"} Oct 03 16:40:23 crc kubenswrapper[4744]: I1003 16:40:23.810600 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-d94d76459-krksv" event={"ID":"df419359-3ee4-4c00-bb37-c7387b19492a","Type":"ContainerStarted","Data":"3ee037de5c91646eaaf0135a3689f32226289716277b10be0ddc093a1f861ef3"} Oct 03 16:40:23 crc kubenswrapper[4744]: I1003 16:40:23.811008 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-d94d76459-krksv" Oct 03 16:40:23 crc kubenswrapper[4744]: I1003 16:40:23.860472 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-d94d76459-krksv" podStartSLOduration=1.927180852 podStartE2EDuration="7.860441991s" podCreationTimestamp="2025-10-03 16:40:16 +0000 UTC" firstStartedPulling="2025-10-03 16:40:17.254857103 +0000 UTC m=+943.534733009" lastFinishedPulling="2025-10-03 16:40:23.188118212 +0000 UTC m=+949.467994148" observedRunningTime="2025-10-03 16:40:23.855478434 +0000 UTC m=+950.135354360" watchObservedRunningTime="2025-10-03 16:40:23.860441991 +0000 UTC m=+950.140317937" Oct 03 16:40:26 crc kubenswrapper[4744]: I1003 16:40:26.796229 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-d94d76459-krksv" Oct 03 16:40:34 crc kubenswrapper[4744]: I1003 16:40:34.668452 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:40:34 crc kubenswrapper[4744]: I1003 16:40:34.669121 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.151261 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6c675fb79f-4st8l"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.153277 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-4st8l" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.156468 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-lfjbc" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.167686 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79d68d6c85-4r792"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.169071 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-4r792" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.169540 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6c675fb79f-4st8l"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.171566 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-js2r8" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.181927 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-bs7f6"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.186547 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-bs7f6" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.193309 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79d68d6c85-4r792"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.193846 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-zrztc" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.211882 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-bs7f6"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.216634 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-846dff85b5-7qtkn"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.217692 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-7qtkn" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.222842 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-sz4dq" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.235064 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-846dff85b5-7qtkn"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.256553 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-599898f689-fxkn9"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.257631 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-599898f689-fxkn9" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.261764 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zbdx\" (UniqueName: \"kubernetes.io/projected/d68c3557-6655-4449-8167-c28d1ccf442e-kube-api-access-4zbdx\") pod \"barbican-operator-controller-manager-6c675fb79f-4st8l\" (UID: \"d68c3557-6655-4449-8167-c28d1ccf442e\") " pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-4st8l" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.261834 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7nm7\" (UniqueName: \"kubernetes.io/projected/a38be30f-e956-4859-9610-a4d0735fe9d4-kube-api-access-h7nm7\") pod \"glance-operator-controller-manager-846dff85b5-7qtkn\" (UID: \"a38be30f-e956-4859-9610-a4d0735fe9d4\") " pod="openstack-operators/glance-operator-controller-manager-846dff85b5-7qtkn" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.262034 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxhrx\" (UniqueName: \"kubernetes.io/projected/da0c0550-089f-4bd9-aaf8-960fcc75f019-kube-api-access-xxhrx\") pod \"heat-operator-controller-manager-599898f689-fxkn9\" (UID: \"da0c0550-089f-4bd9-aaf8-960fcc75f019\") " pod="openstack-operators/heat-operator-controller-manager-599898f689-fxkn9" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.262084 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69qhn\" (UniqueName: \"kubernetes.io/projected/829ba77b-d3cc-416c-8813-eeff7b6b60bd-kube-api-access-69qhn\") pod \"cinder-operator-controller-manager-79d68d6c85-4r792\" (UID: \"829ba77b-d3cc-416c-8813-eeff7b6b60bd\") " pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-4r792" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.262173 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqhk4\" (UniqueName: \"kubernetes.io/projected/19371562-3c8d-4dce-864c-0a37cbce2bf3-kube-api-access-sqhk4\") pod \"designate-operator-controller-manager-75dfd9b554-bs7f6\" (UID: \"19371562-3c8d-4dce-864c-0a37cbce2bf3\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-bs7f6" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.264101 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-qqwcq" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.265734 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-599898f689-fxkn9"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.275488 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6769b867d9-n26jb"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.276510 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-n26jb" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.284462 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-skvhq" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.293655 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.294647 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.297753 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.298086 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-9qwf5" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.307034 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6769b867d9-n26jb"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.318846 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.324308 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-84bc9db6cc-tb5k2"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.325329 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-tb5k2" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.329411 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-bsqnh" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.355551 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-84bc9db6cc-tb5k2"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.366093 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.382262 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zbdx\" (UniqueName: \"kubernetes.io/projected/d68c3557-6655-4449-8167-c28d1ccf442e-kube-api-access-4zbdx\") pod \"barbican-operator-controller-manager-6c675fb79f-4st8l\" (UID: \"d68c3557-6655-4449-8167-c28d1ccf442e\") " pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-4st8l" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.382340 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7nm7\" (UniqueName: \"kubernetes.io/projected/a38be30f-e956-4859-9610-a4d0735fe9d4-kube-api-access-h7nm7\") pod \"glance-operator-controller-manager-846dff85b5-7qtkn\" (UID: \"a38be30f-e956-4859-9610-a4d0735fe9d4\") " pod="openstack-operators/glance-operator-controller-manager-846dff85b5-7qtkn" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.382606 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxhrx\" (UniqueName: \"kubernetes.io/projected/da0c0550-089f-4bd9-aaf8-960fcc75f019-kube-api-access-xxhrx\") pod \"heat-operator-controller-manager-599898f689-fxkn9\" (UID: \"da0c0550-089f-4bd9-aaf8-960fcc75f019\") " pod="openstack-operators/heat-operator-controller-manager-599898f689-fxkn9" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.382687 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69qhn\" (UniqueName: \"kubernetes.io/projected/829ba77b-d3cc-416c-8813-eeff7b6b60bd-kube-api-access-69qhn\") pod \"cinder-operator-controller-manager-79d68d6c85-4r792\" (UID: \"829ba77b-d3cc-416c-8813-eeff7b6b60bd\") " pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-4r792" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.382859 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqhk4\" (UniqueName: \"kubernetes.io/projected/19371562-3c8d-4dce-864c-0a37cbce2bf3-kube-api-access-sqhk4\") pod \"designate-operator-controller-manager-75dfd9b554-bs7f6\" (UID: \"19371562-3c8d-4dce-864c-0a37cbce2bf3\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-bs7f6" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.384860 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.411323 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-m8rdn" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.425476 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6fd6854b49-zd5fc"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.427252 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-zd5fc" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.444712 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-djrrh" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.450696 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.461198 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.461993 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zbdx\" (UniqueName: \"kubernetes.io/projected/d68c3557-6655-4449-8167-c28d1ccf442e-kube-api-access-4zbdx\") pod \"barbican-operator-controller-manager-6c675fb79f-4st8l\" (UID: \"d68c3557-6655-4449-8167-c28d1ccf442e\") " pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-4st8l" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.462604 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.463170 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxhrx\" (UniqueName: \"kubernetes.io/projected/da0c0550-089f-4bd9-aaf8-960fcc75f019-kube-api-access-xxhrx\") pod \"heat-operator-controller-manager-599898f689-fxkn9\" (UID: \"da0c0550-089f-4bd9-aaf8-960fcc75f019\") " pod="openstack-operators/heat-operator-controller-manager-599898f689-fxkn9" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.463282 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7nm7\" (UniqueName: \"kubernetes.io/projected/a38be30f-e956-4859-9610-a4d0735fe9d4-kube-api-access-h7nm7\") pod \"glance-operator-controller-manager-846dff85b5-7qtkn\" (UID: \"a38be30f-e956-4859-9610-a4d0735fe9d4\") " pod="openstack-operators/glance-operator-controller-manager-846dff85b5-7qtkn" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.465332 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqhk4\" (UniqueName: \"kubernetes.io/projected/19371562-3c8d-4dce-864c-0a37cbce2bf3-kube-api-access-sqhk4\") pod \"designate-operator-controller-manager-75dfd9b554-bs7f6\" (UID: \"19371562-3c8d-4dce-864c-0a37cbce2bf3\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-bs7f6" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.466288 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69qhn\" (UniqueName: \"kubernetes.io/projected/829ba77b-d3cc-416c-8813-eeff7b6b60bd-kube-api-access-69qhn\") pod \"cinder-operator-controller-manager-79d68d6c85-4r792\" (UID: \"829ba77b-d3cc-416c-8813-eeff7b6b60bd\") " pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-4r792" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.472050 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-qlvbh" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.473739 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-4st8l" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.487679 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3bc36b86-4888-4341-b52d-b9b73013f9ad-cert\") pod \"infra-operator-controller-manager-5fbf469cd7-4jdmt\" (UID: \"3bc36b86-4888-4341-b52d-b9b73013f9ad\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.487746 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8gs9\" (UniqueName: \"kubernetes.io/projected/23519c20-6aec-4461-bea9-adaf88e87ab3-kube-api-access-b8gs9\") pod \"ironic-operator-controller-manager-84bc9db6cc-tb5k2\" (UID: \"23519c20-6aec-4461-bea9-adaf88e87ab3\") " pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-tb5k2" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.487768 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45nmf\" (UniqueName: \"kubernetes.io/projected/3bc36b86-4888-4341-b52d-b9b73013f9ad-kube-api-access-45nmf\") pod \"infra-operator-controller-manager-5fbf469cd7-4jdmt\" (UID: \"3bc36b86-4888-4341-b52d-b9b73013f9ad\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.487871 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5r7b\" (UniqueName: \"kubernetes.io/projected/fd8ecb24-30b8-4012-846a-4618f178996b-kube-api-access-k5r7b\") pod \"horizon-operator-controller-manager-6769b867d9-n26jb\" (UID: \"fd8ecb24-30b8-4012-846a-4618f178996b\") " pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-n26jb" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.487915 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59p5n\" (UniqueName: \"kubernetes.io/projected/a862283f-4243-42bd-bb2c-5b2392d657ac-kube-api-access-59p5n\") pod \"keystone-operator-controller-manager-7f55849f88-r8tdt\" (UID: \"a862283f-4243-42bd-bb2c-5b2392d657ac\") " pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.491989 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-4r792" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.502892 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6fd6854b49-zd5fc"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.516632 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-bs7f6" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.530814 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.537291 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-7qtkn" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.554719 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6574bf987d-xxdpr"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.556072 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-xxdpr" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.559802 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-79rhh" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.576661 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-599898f689-fxkn9" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.589139 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59p5n\" (UniqueName: \"kubernetes.io/projected/a862283f-4243-42bd-bb2c-5b2392d657ac-kube-api-access-59p5n\") pod \"keystone-operator-controller-manager-7f55849f88-r8tdt\" (UID: \"a862283f-4243-42bd-bb2c-5b2392d657ac\") " pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.589203 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3bc36b86-4888-4341-b52d-b9b73013f9ad-cert\") pod \"infra-operator-controller-manager-5fbf469cd7-4jdmt\" (UID: \"3bc36b86-4888-4341-b52d-b9b73013f9ad\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.589231 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8gs9\" (UniqueName: \"kubernetes.io/projected/23519c20-6aec-4461-bea9-adaf88e87ab3-kube-api-access-b8gs9\") pod \"ironic-operator-controller-manager-84bc9db6cc-tb5k2\" (UID: \"23519c20-6aec-4461-bea9-adaf88e87ab3\") " pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-tb5k2" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.589250 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45nmf\" (UniqueName: \"kubernetes.io/projected/3bc36b86-4888-4341-b52d-b9b73013f9ad-kube-api-access-45nmf\") pod \"infra-operator-controller-manager-5fbf469cd7-4jdmt\" (UID: \"3bc36b86-4888-4341-b52d-b9b73013f9ad\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.589305 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kfbj\" (UniqueName: \"kubernetes.io/projected/be9443a6-46c8-4374-8caa-8040ae117981-kube-api-access-8kfbj\") pod \"manila-operator-controller-manager-6fd6854b49-zd5fc\" (UID: \"be9443a6-46c8-4374-8caa-8040ae117981\") " pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-zd5fc" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.589329 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5r7b\" (UniqueName: \"kubernetes.io/projected/fd8ecb24-30b8-4012-846a-4618f178996b-kube-api-access-k5r7b\") pod \"horizon-operator-controller-manager-6769b867d9-n26jb\" (UID: \"fd8ecb24-30b8-4012-846a-4618f178996b\") " pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-n26jb" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.589359 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phtms\" (UniqueName: \"kubernetes.io/projected/d34995fe-607b-46ae-a2eb-c91d9c244415-kube-api-access-phtms\") pod \"mariadb-operator-controller-manager-5c468bf4d4-kzsg2\" (UID: \"d34995fe-607b-46ae-a2eb-c91d9c244415\") " pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2" Oct 03 16:40:44 crc kubenswrapper[4744]: E1003 16:40:44.589787 4744 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 03 16:40:44 crc kubenswrapper[4744]: E1003 16:40:44.589831 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3bc36b86-4888-4341-b52d-b9b73013f9ad-cert podName:3bc36b86-4888-4341-b52d-b9b73013f9ad nodeName:}" failed. No retries permitted until 2025-10-03 16:40:45.08981364 +0000 UTC m=+971.369689536 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3bc36b86-4888-4341-b52d-b9b73013f9ad-cert") pod "infra-operator-controller-manager-5fbf469cd7-4jdmt" (UID: "3bc36b86-4888-4341-b52d-b9b73013f9ad") : secret "infra-operator-webhook-server-cert" not found Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.608428 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59p5n\" (UniqueName: \"kubernetes.io/projected/a862283f-4243-42bd-bb2c-5b2392d657ac-kube-api-access-59p5n\") pod \"keystone-operator-controller-manager-7f55849f88-r8tdt\" (UID: \"a862283f-4243-42bd-bb2c-5b2392d657ac\") " pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.608431 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45nmf\" (UniqueName: \"kubernetes.io/projected/3bc36b86-4888-4341-b52d-b9b73013f9ad-kube-api-access-45nmf\") pod \"infra-operator-controller-manager-5fbf469cd7-4jdmt\" (UID: \"3bc36b86-4888-4341-b52d-b9b73013f9ad\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.609656 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6574bf987d-xxdpr"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.609940 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8gs9\" (UniqueName: \"kubernetes.io/projected/23519c20-6aec-4461-bea9-adaf88e87ab3-kube-api-access-b8gs9\") pod \"ironic-operator-controller-manager-84bc9db6cc-tb5k2\" (UID: \"23519c20-6aec-4461-bea9-adaf88e87ab3\") " pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-tb5k2" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.611143 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5r7b\" (UniqueName: \"kubernetes.io/projected/fd8ecb24-30b8-4012-846a-4618f178996b-kube-api-access-k5r7b\") pod \"horizon-operator-controller-manager-6769b867d9-n26jb\" (UID: \"fd8ecb24-30b8-4012-846a-4618f178996b\") " pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-n26jb" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.617672 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.622104 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.627512 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-rph7w" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.644993 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-59d6cfdf45-xr6db"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.645313 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-tb5k2" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.646446 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-xr6db" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.653802 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-h946r" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.662432 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.689553 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-59d6cfdf45-xr6db"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.690267 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kfbj\" (UniqueName: \"kubernetes.io/projected/be9443a6-46c8-4374-8caa-8040ae117981-kube-api-access-8kfbj\") pod \"manila-operator-controller-manager-6fd6854b49-zd5fc\" (UID: \"be9443a6-46c8-4374-8caa-8040ae117981\") " pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-zd5fc" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.690308 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phtms\" (UniqueName: \"kubernetes.io/projected/d34995fe-607b-46ae-a2eb-c91d9c244415-kube-api-access-phtms\") pod \"mariadb-operator-controller-manager-5c468bf4d4-kzsg2\" (UID: \"d34995fe-607b-46ae-a2eb-c91d9c244415\") " pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.690347 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4gjm\" (UniqueName: \"kubernetes.io/projected/2832b841-b383-4eed-a8ce-71062269238a-kube-api-access-l4gjm\") pod \"neutron-operator-controller-manager-6574bf987d-xxdpr\" (UID: \"2832b841-b383-4eed-a8ce-71062269238a\") " pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-xxdpr" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.694426 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.695894 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.697431 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-sp6zz" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.703557 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-688db7b6c7-jhqdv"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.704953 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-jhqdv" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.706642 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.707153 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-6trl9" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.724707 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phtms\" (UniqueName: \"kubernetes.io/projected/d34995fe-607b-46ae-a2eb-c91d9c244415-kube-api-access-phtms\") pod \"mariadb-operator-controller-manager-5c468bf4d4-kzsg2\" (UID: \"d34995fe-607b-46ae-a2eb-c91d9c244415\") " pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.725600 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.743380 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-7d8bb7f44c-jktv9"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.743834 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kfbj\" (UniqueName: \"kubernetes.io/projected/be9443a6-46c8-4374-8caa-8040ae117981-kube-api-access-8kfbj\") pod \"manila-operator-controller-manager-6fd6854b49-zd5fc\" (UID: \"be9443a6-46c8-4374-8caa-8040ae117981\") " pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-zd5fc" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.744755 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-jktv9" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.748409 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-82wz7" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.754614 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-688db7b6c7-jhqdv"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.773145 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.784201 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-psqc4"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.785892 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-psqc4" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.792918 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62bjh\" (UniqueName: \"kubernetes.io/projected/911d5ffd-d447-4b8a-a177-7fc738d2aa35-kube-api-access-62bjh\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678v879q\" (UID: \"911d5ffd-d447-4b8a-a177-7fc738d2aa35\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.793173 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4gjm\" (UniqueName: \"kubernetes.io/projected/2832b841-b383-4eed-a8ce-71062269238a-kube-api-access-l4gjm\") pod \"neutron-operator-controller-manager-6574bf987d-xxdpr\" (UID: \"2832b841-b383-4eed-a8ce-71062269238a\") " pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-xxdpr" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.793630 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/911d5ffd-d447-4b8a-a177-7fc738d2aa35-cert\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678v879q\" (UID: \"911d5ffd-d447-4b8a-a177-7fc738d2aa35\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.794115 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbwpl\" (UniqueName: \"kubernetes.io/projected/8be63e93-5f9d-4ffe-9ab2-d5b794c34f9b-kube-api-access-xbwpl\") pod \"ovn-operator-controller-manager-688db7b6c7-jhqdv\" (UID: \"8be63e93-5f9d-4ffe-9ab2-d5b794c34f9b\") " pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-jhqdv" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.794763 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47z7h\" (UniqueName: \"kubernetes.io/projected/d34377f2-1b67-4c62-9602-ddb93f0cb4d7-kube-api-access-47z7h\") pod \"octavia-operator-controller-manager-59d6cfdf45-xr6db\" (UID: \"d34377f2-1b67-4c62-9602-ddb93f0cb4d7\") " pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-xr6db" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.795262 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc5wh\" (UniqueName: \"kubernetes.io/projected/438b2904-4b85-4d82-b883-f5788d4caab9-kube-api-access-zc5wh\") pod \"nova-operator-controller-manager-555c7456bd-6xg8v\" (UID: \"438b2904-4b85-4d82-b883-f5788d4caab9\") " pod="openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.795718 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-jhv7s" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.800058 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-7d8bb7f44c-jktv9"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.816061 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-psqc4"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.823958 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.830790 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.834302 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-lrp5m" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.842128 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4gjm\" (UniqueName: \"kubernetes.io/projected/2832b841-b383-4eed-a8ce-71062269238a-kube-api-access-l4gjm\") pod \"neutron-operator-controller-manager-6574bf987d-xxdpr\" (UID: \"2832b841-b383-4eed-a8ce-71062269238a\") " pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-xxdpr" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.847627 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-zd5fc" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.859683 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.878973 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.893065 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.895780 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.900916 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtdh2\" (UniqueName: \"kubernetes.io/projected/a6943fb4-ab7a-445f-b4d1-be9089562b77-kube-api-access-xtdh2\") pod \"swift-operator-controller-manager-6859f9b676-psqc4\" (UID: \"a6943fb4-ab7a-445f-b4d1-be9089562b77\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-psqc4" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.900964 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtfn6\" (UniqueName: \"kubernetes.io/projected/39de7155-617f-444a-bb16-0cb7fb229eb7-kube-api-access-mtfn6\") pod \"placement-operator-controller-manager-7d8bb7f44c-jktv9\" (UID: \"39de7155-617f-444a-bb16-0cb7fb229eb7\") " pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-jktv9" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.901013 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc5wh\" (UniqueName: \"kubernetes.io/projected/438b2904-4b85-4d82-b883-f5788d4caab9-kube-api-access-zc5wh\") pod \"nova-operator-controller-manager-555c7456bd-6xg8v\" (UID: \"438b2904-4b85-4d82-b883-f5788d4caab9\") " pod="openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.901074 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62bjh\" (UniqueName: \"kubernetes.io/projected/911d5ffd-d447-4b8a-a177-7fc738d2aa35-kube-api-access-62bjh\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678v879q\" (UID: \"911d5ffd-d447-4b8a-a177-7fc738d2aa35\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.901095 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/911d5ffd-d447-4b8a-a177-7fc738d2aa35-cert\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678v879q\" (UID: \"911d5ffd-d447-4b8a-a177-7fc738d2aa35\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.901143 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbwpl\" (UniqueName: \"kubernetes.io/projected/8be63e93-5f9d-4ffe-9ab2-d5b794c34f9b-kube-api-access-xbwpl\") pod \"ovn-operator-controller-manager-688db7b6c7-jhqdv\" (UID: \"8be63e93-5f9d-4ffe-9ab2-d5b794c34f9b\") " pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-jhqdv" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.901172 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47z7h\" (UniqueName: \"kubernetes.io/projected/d34377f2-1b67-4c62-9602-ddb93f0cb4d7-kube-api-access-47z7h\") pod \"octavia-operator-controller-manager-59d6cfdf45-xr6db\" (UID: \"d34377f2-1b67-4c62-9602-ddb93f0cb4d7\") " pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-xr6db" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.900942 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-n26jb" Oct 03 16:40:44 crc kubenswrapper[4744]: E1003 16:40:44.901575 4744 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 03 16:40:44 crc kubenswrapper[4744]: E1003 16:40:44.901622 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/911d5ffd-d447-4b8a-a177-7fc738d2aa35-cert podName:911d5ffd-d447-4b8a-a177-7fc738d2aa35 nodeName:}" failed. No retries permitted until 2025-10-03 16:40:45.401606004 +0000 UTC m=+971.681481900 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/911d5ffd-d447-4b8a-a177-7fc738d2aa35-cert") pod "openstack-baremetal-operator-controller-manager-6f64c4d678v879q" (UID: "911d5ffd-d447-4b8a-a177-7fc738d2aa35") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.907714 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-xwwjq" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.908563 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-xxdpr" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.953247 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89"] Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.981863 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbwpl\" (UniqueName: \"kubernetes.io/projected/8be63e93-5f9d-4ffe-9ab2-d5b794c34f9b-kube-api-access-xbwpl\") pod \"ovn-operator-controller-manager-688db7b6c7-jhqdv\" (UID: \"8be63e93-5f9d-4ffe-9ab2-d5b794c34f9b\") " pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-jhqdv" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.987332 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47z7h\" (UniqueName: \"kubernetes.io/projected/d34377f2-1b67-4c62-9602-ddb93f0cb4d7-kube-api-access-47z7h\") pod \"octavia-operator-controller-manager-59d6cfdf45-xr6db\" (UID: \"d34377f2-1b67-4c62-9602-ddb93f0cb4d7\") " pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-xr6db" Oct 03 16:40:44 crc kubenswrapper[4744]: I1003 16:40:44.992046 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc5wh\" (UniqueName: \"kubernetes.io/projected/438b2904-4b85-4d82-b883-f5788d4caab9-kube-api-access-zc5wh\") pod \"nova-operator-controller-manager-555c7456bd-6xg8v\" (UID: \"438b2904-4b85-4d82-b883-f5788d4caab9\") " pod="openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.000330 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62bjh\" (UniqueName: \"kubernetes.io/projected/911d5ffd-d447-4b8a-a177-7fc738d2aa35-kube-api-access-62bjh\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678v879q\" (UID: \"911d5ffd-d447-4b8a-a177-7fc738d2aa35\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.009592 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-fcd7d9895-wjl4s"] Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.010627 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wqwq\" (UniqueName: \"kubernetes.io/projected/09c1f75f-d699-4675-b78d-a37353acd758-kube-api-access-7wqwq\") pod \"telemetry-operator-controller-manager-5db5cf686f-7qgsm\" (UID: \"09c1f75f-d699-4675-b78d-a37353acd758\") " pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.010874 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtdh2\" (UniqueName: \"kubernetes.io/projected/a6943fb4-ab7a-445f-b4d1-be9089562b77-kube-api-access-xtdh2\") pod \"swift-operator-controller-manager-6859f9b676-psqc4\" (UID: \"a6943fb4-ab7a-445f-b4d1-be9089562b77\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-psqc4" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.010974 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtfn6\" (UniqueName: \"kubernetes.io/projected/39de7155-617f-444a-bb16-0cb7fb229eb7-kube-api-access-mtfn6\") pod \"placement-operator-controller-manager-7d8bb7f44c-jktv9\" (UID: \"39de7155-617f-444a-bb16-0cb7fb229eb7\") " pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-jktv9" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.011059 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tvfd\" (UniqueName: \"kubernetes.io/projected/1d753e81-b04a-4216-991f-256a2ab6ac94-kube-api-access-7tvfd\") pod \"test-operator-controller-manager-5cd5cb47d7-ksw89\" (UID: \"1d753e81-b04a-4216-991f-256a2ab6ac94\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.015365 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-wjl4s" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.024984 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-fcfvh" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.048004 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-fcd7d9895-wjl4s"] Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.058446 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtdh2\" (UniqueName: \"kubernetes.io/projected/a6943fb4-ab7a-445f-b4d1-be9089562b77-kube-api-access-xtdh2\") pod \"swift-operator-controller-manager-6859f9b676-psqc4\" (UID: \"a6943fb4-ab7a-445f-b4d1-be9089562b77\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-psqc4" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.062418 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtfn6\" (UniqueName: \"kubernetes.io/projected/39de7155-617f-444a-bb16-0cb7fb229eb7-kube-api-access-mtfn6\") pod \"placement-operator-controller-manager-7d8bb7f44c-jktv9\" (UID: \"39de7155-617f-444a-bb16-0cb7fb229eb7\") " pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-jktv9" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.071008 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl"] Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.072770 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.079710 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl"] Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.082113 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-whkhs" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.082268 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.094609 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw"] Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.096029 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw"] Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.096134 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.096154 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-xr6db" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.098545 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-qv7rt" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.112525 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3bc36b86-4888-4341-b52d-b9b73013f9ad-cert\") pod \"infra-operator-controller-manager-5fbf469cd7-4jdmt\" (UID: \"3bc36b86-4888-4341-b52d-b9b73013f9ad\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.112611 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wqwq\" (UniqueName: \"kubernetes.io/projected/09c1f75f-d699-4675-b78d-a37353acd758-kube-api-access-7wqwq\") pod \"telemetry-operator-controller-manager-5db5cf686f-7qgsm\" (UID: \"09c1f75f-d699-4675-b78d-a37353acd758\") " pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.112653 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tvfd\" (UniqueName: \"kubernetes.io/projected/1d753e81-b04a-4216-991f-256a2ab6ac94-kube-api-access-7tvfd\") pod \"test-operator-controller-manager-5cd5cb47d7-ksw89\" (UID: \"1d753e81-b04a-4216-991f-256a2ab6ac94\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.112682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p8nw\" (UniqueName: \"kubernetes.io/projected/1aa4d11e-aebd-477b-98e1-a05ea74a7677-kube-api-access-8p8nw\") pod \"watcher-operator-controller-manager-fcd7d9895-wjl4s\" (UID: \"1aa4d11e-aebd-477b-98e1-a05ea74a7677\") " pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-wjl4s" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.127677 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3bc36b86-4888-4341-b52d-b9b73013f9ad-cert\") pod \"infra-operator-controller-manager-5fbf469cd7-4jdmt\" (UID: \"3bc36b86-4888-4341-b52d-b9b73013f9ad\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.136669 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tvfd\" (UniqueName: \"kubernetes.io/projected/1d753e81-b04a-4216-991f-256a2ab6ac94-kube-api-access-7tvfd\") pod \"test-operator-controller-manager-5cd5cb47d7-ksw89\" (UID: \"1d753e81-b04a-4216-991f-256a2ab6ac94\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.143650 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wqwq\" (UniqueName: \"kubernetes.io/projected/09c1f75f-d699-4675-b78d-a37353acd758-kube-api-access-7wqwq\") pod \"telemetry-operator-controller-manager-5db5cf686f-7qgsm\" (UID: \"09c1f75f-d699-4675-b78d-a37353acd758\") " pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.151456 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-jhqdv" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.157975 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6c675fb79f-4st8l"] Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.161763 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-jktv9" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.177095 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-psqc4" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.192988 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.213268 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.213695 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stkwt\" (UniqueName: \"kubernetes.io/projected/4717fade-30fb-4f78-b7fb-dec2799d9b8c-kube-api-access-stkwt\") pod \"openstack-operator-controller-manager-7758d4b55d-6xczl\" (UID: \"4717fade-30fb-4f78-b7fb-dec2799d9b8c\") " pod="openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.213762 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4717fade-30fb-4f78-b7fb-dec2799d9b8c-cert\") pod \"openstack-operator-controller-manager-7758d4b55d-6xczl\" (UID: \"4717fade-30fb-4f78-b7fb-dec2799d9b8c\") " pod="openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.213809 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhvnw\" (UniqueName: \"kubernetes.io/projected/aef71e84-6035-42fb-a16f-c0e97876f1c0-kube-api-access-nhvnw\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw\" (UID: \"aef71e84-6035-42fb-a16f-c0e97876f1c0\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.213860 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p8nw\" (UniqueName: \"kubernetes.io/projected/1aa4d11e-aebd-477b-98e1-a05ea74a7677-kube-api-access-8p8nw\") pod \"watcher-operator-controller-manager-fcd7d9895-wjl4s\" (UID: \"1aa4d11e-aebd-477b-98e1-a05ea74a7677\") " pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-wjl4s" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.221981 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.232525 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p8nw\" (UniqueName: \"kubernetes.io/projected/1aa4d11e-aebd-477b-98e1-a05ea74a7677-kube-api-access-8p8nw\") pod \"watcher-operator-controller-manager-fcd7d9895-wjl4s\" (UID: \"1aa4d11e-aebd-477b-98e1-a05ea74a7677\") " pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-wjl4s" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.272820 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v" Oct 03 16:40:45 crc kubenswrapper[4744]: W1003 16:40:45.295777 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd68c3557_6655_4449_8167_c28d1ccf442e.slice/crio-b2daad57a651e1264b8918708c2df24d5a8d3be7477c208866631c80320a4bce WatchSource:0}: Error finding container b2daad57a651e1264b8918708c2df24d5a8d3be7477c208866631c80320a4bce: Status 404 returned error can't find the container with id b2daad57a651e1264b8918708c2df24d5a8d3be7477c208866631c80320a4bce Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.314970 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhvnw\" (UniqueName: \"kubernetes.io/projected/aef71e84-6035-42fb-a16f-c0e97876f1c0-kube-api-access-nhvnw\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw\" (UID: \"aef71e84-6035-42fb-a16f-c0e97876f1c0\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.315068 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stkwt\" (UniqueName: \"kubernetes.io/projected/4717fade-30fb-4f78-b7fb-dec2799d9b8c-kube-api-access-stkwt\") pod \"openstack-operator-controller-manager-7758d4b55d-6xczl\" (UID: \"4717fade-30fb-4f78-b7fb-dec2799d9b8c\") " pod="openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.315156 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4717fade-30fb-4f78-b7fb-dec2799d9b8c-cert\") pod \"openstack-operator-controller-manager-7758d4b55d-6xczl\" (UID: \"4717fade-30fb-4f78-b7fb-dec2799d9b8c\") " pod="openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.321890 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4717fade-30fb-4f78-b7fb-dec2799d9b8c-cert\") pod \"openstack-operator-controller-manager-7758d4b55d-6xczl\" (UID: \"4717fade-30fb-4f78-b7fb-dec2799d9b8c\") " pod="openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.338228 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhvnw\" (UniqueName: \"kubernetes.io/projected/aef71e84-6035-42fb-a16f-c0e97876f1c0-kube-api-access-nhvnw\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw\" (UID: \"aef71e84-6035-42fb-a16f-c0e97876f1c0\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.339257 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stkwt\" (UniqueName: \"kubernetes.io/projected/4717fade-30fb-4f78-b7fb-dec2799d9b8c-kube-api-access-stkwt\") pod \"openstack-operator-controller-manager-7758d4b55d-6xczl\" (UID: \"4717fade-30fb-4f78-b7fb-dec2799d9b8c\") " pod="openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.425601 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/911d5ffd-d447-4b8a-a177-7fc738d2aa35-cert\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678v879q\" (UID: \"911d5ffd-d447-4b8a-a177-7fc738d2aa35\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" Oct 03 16:40:45 crc kubenswrapper[4744]: E1003 16:40:45.425853 4744 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 03 16:40:45 crc kubenswrapper[4744]: E1003 16:40:45.425943 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/911d5ffd-d447-4b8a-a177-7fc738d2aa35-cert podName:911d5ffd-d447-4b8a-a177-7fc738d2aa35 nodeName:}" failed. No retries permitted until 2025-10-03 16:40:46.425921013 +0000 UTC m=+972.705796909 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/911d5ffd-d447-4b8a-a177-7fc738d2aa35-cert") pod "openstack-baremetal-operator-controller-manager-6f64c4d678v879q" (UID: "911d5ffd-d447-4b8a-a177-7fc738d2aa35") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.514842 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-wjl4s" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.530757 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.539147 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79d68d6c85-4r792"] Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.540247 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw" Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.854203 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-bs7f6"] Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.881565 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt"] Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.893042 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-84bc9db6cc-tb5k2"] Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.896627 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-846dff85b5-7qtkn"] Oct 03 16:40:45 crc kubenswrapper[4744]: W1003 16:40:45.900532 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda38be30f_e956_4859_9610_a4d0735fe9d4.slice/crio-d4c7a48370a4579872c6251f0f85d940832e6bdd1168b4dd0d1a9f551b8eb286 WatchSource:0}: Error finding container d4c7a48370a4579872c6251f0f85d940832e6bdd1168b4dd0d1a9f551b8eb286: Status 404 returned error can't find the container with id d4c7a48370a4579872c6251f0f85d940832e6bdd1168b4dd0d1a9f551b8eb286 Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.916791 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-599898f689-fxkn9"] Oct 03 16:40:45 crc kubenswrapper[4744]: W1003 16:40:45.952450 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda0c0550_089f_4bd9_aaf8_960fcc75f019.slice/crio-bfca3d0f7008b0fd70eef8a6b6fdd995460359c97e742e39e5d5974eb34d0114 WatchSource:0}: Error finding container bfca3d0f7008b0fd70eef8a6b6fdd995460359c97e742e39e5d5974eb34d0114: Status 404 returned error can't find the container with id bfca3d0f7008b0fd70eef8a6b6fdd995460359c97e742e39e5d5974eb34d0114 Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.953301 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-688db7b6c7-jhqdv"] Oct 03 16:40:45 crc kubenswrapper[4744]: W1003 16:40:45.955439 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8be63e93_5f9d_4ffe_9ab2_d5b794c34f9b.slice/crio-3fd6e1c098bfaf69fe96d0381262d043a74449d7eb62b618c746eac10e4f1f9d WatchSource:0}: Error finding container 3fd6e1c098bfaf69fe96d0381262d043a74449d7eb62b618c746eac10e4f1f9d: Status 404 returned error can't find the container with id 3fd6e1c098bfaf69fe96d0381262d043a74449d7eb62b618c746eac10e4f1f9d Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.961476 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6fd6854b49-zd5fc"] Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.967015 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6574bf987d-xxdpr"] Oct 03 16:40:45 crc kubenswrapper[4744]: W1003 16:40:45.970831 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd34377f2_1b67_4c62_9602_ddb93f0cb4d7.slice/crio-d00e8965468f96966203b40784043c2ab35779f06472ae33840417950d4d4b10 WatchSource:0}: Error finding container d00e8965468f96966203b40784043c2ab35779f06472ae33840417950d4d4b10: Status 404 returned error can't find the container with id d00e8965468f96966203b40784043c2ab35779f06472ae33840417950d4d4b10 Oct 03 16:40:45 crc kubenswrapper[4744]: W1003 16:40:45.973625 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd8ecb24_30b8_4012_846a_4618f178996b.slice/crio-0cc3eb82d55d4855df84a8b4f9f08506ac42d7d37459e0894df11f801e8e7148 WatchSource:0}: Error finding container 0cc3eb82d55d4855df84a8b4f9f08506ac42d7d37459e0894df11f801e8e7148: Status 404 returned error can't find the container with id 0cc3eb82d55d4855df84a8b4f9f08506ac42d7d37459e0894df11f801e8e7148 Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.974997 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-59d6cfdf45-xr6db"] Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.980291 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6769b867d9-n26jb"] Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.995466 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-7qtkn" event={"ID":"a38be30f-e956-4859-9610-a4d0735fe9d4","Type":"ContainerStarted","Data":"d4c7a48370a4579872c6251f0f85d940832e6bdd1168b4dd0d1a9f551b8eb286"} Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.996810 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-4st8l" event={"ID":"d68c3557-6655-4449-8167-c28d1ccf442e","Type":"ContainerStarted","Data":"b2daad57a651e1264b8918708c2df24d5a8d3be7477c208866631c80320a4bce"} Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.997962 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-4r792" event={"ID":"829ba77b-d3cc-416c-8813-eeff7b6b60bd","Type":"ContainerStarted","Data":"fddcf768ee71006018b04f380957708a40c49e3779223253d2bcffefffbcbb52"} Oct 03 16:40:45 crc kubenswrapper[4744]: I1003 16:40:45.999049 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-n26jb" event={"ID":"fd8ecb24-30b8-4012-846a-4618f178996b","Type":"ContainerStarted","Data":"0cc3eb82d55d4855df84a8b4f9f08506ac42d7d37459e0894df11f801e8e7148"} Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.000128 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-xxdpr" event={"ID":"2832b841-b383-4eed-a8ce-71062269238a","Type":"ContainerStarted","Data":"a5826477820656006c6ff8cd048a01f03bbd23758864f93a54d569dae2660099"} Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.001162 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-zd5fc" event={"ID":"be9443a6-46c8-4374-8caa-8040ae117981","Type":"ContainerStarted","Data":"ecdf392b38f11c6e6c0cf7ceb15436eece0bad217ddec30c3f0cc8d523f582d0"} Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.002029 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-tb5k2" event={"ID":"23519c20-6aec-4461-bea9-adaf88e87ab3","Type":"ContainerStarted","Data":"a5cdc188bcc44b0978b4972aa6cd7f6c5b61d15ae356b87c6e426fb23b933192"} Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.003123 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-jhqdv" event={"ID":"8be63e93-5f9d-4ffe-9ab2-d5b794c34f9b","Type":"ContainerStarted","Data":"3fd6e1c098bfaf69fe96d0381262d043a74449d7eb62b618c746eac10e4f1f9d"} Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.011402 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt" event={"ID":"a862283f-4243-42bd-bb2c-5b2392d657ac","Type":"ContainerStarted","Data":"9ae09f6884e300db7a79a049531332370b5bb4a61bfb7f77c82a5b4753664b81"} Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.015567 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-xr6db" event={"ID":"d34377f2-1b67-4c62-9602-ddb93f0cb4d7","Type":"ContainerStarted","Data":"d00e8965468f96966203b40784043c2ab35779f06472ae33840417950d4d4b10"} Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.016895 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-bs7f6" event={"ID":"19371562-3c8d-4dce-864c-0a37cbce2bf3","Type":"ContainerStarted","Data":"9c02c61e79aedfb34b52f43dcefc875311a2689f4ae734ac23399ff3ff0faaa1"} Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.018731 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-599898f689-fxkn9" event={"ID":"da0c0550-089f-4bd9-aaf8-960fcc75f019","Type":"ContainerStarted","Data":"bfca3d0f7008b0fd70eef8a6b6fdd995460359c97e742e39e5d5974eb34d0114"} Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.092775 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-7d8bb7f44c-jktv9"] Oct 03 16:40:46 crc kubenswrapper[4744]: W1003 16:40:46.103436 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39de7155_617f_444a_bb16_0cb7fb229eb7.slice/crio-c1937ea9665a6039d0d0eb996964e234baf94c1b3621996b68d5cb27d8d72e17 WatchSource:0}: Error finding container c1937ea9665a6039d0d0eb996964e234baf94c1b3621996b68d5cb27d8d72e17: Status 404 returned error can't find the container with id c1937ea9665a6039d0d0eb996964e234baf94c1b3621996b68d5cb27d8d72e17 Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.106570 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt"] Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.128011 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm"] Oct 03 16:40:46 crc kubenswrapper[4744]: E1003 16:40:46.136339 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:40fb1819b6639807b77ef79448d35f1e4bfc1838a09d4f380e9fa0f755352475,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-45nmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-5fbf469cd7-4jdmt_openstack-operators(3bc36b86-4888-4341-b52d-b9b73013f9ad): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.140548 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2"] Oct 03 16:40:46 crc kubenswrapper[4744]: E1003 16:40:46.146723 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:8f5eee2eb7b77432ef1a88ed693ff981514359dfc808581f393bcef252de5cfa,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7wqwq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5db5cf686f-7qgsm_openstack-operators(09c1f75f-d699-4675-b78d-a37353acd758): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 16:40:46 crc kubenswrapper[4744]: E1003 16:40:46.156598 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:110b885fe640ffdd8536e7da2a613677a6777e3d902e2ff15fa4d5968fe06c54,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-phtms,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-5c468bf4d4-kzsg2_openstack-operators(d34995fe-607b-46ae-a2eb-c91d9c244415): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 16:40:46 crc kubenswrapper[4744]: E1003 16:40:46.331338 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm" podUID="09c1f75f-d699-4675-b78d-a37353acd758" Oct 03 16:40:46 crc kubenswrapper[4744]: E1003 16:40:46.340303 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" podUID="3bc36b86-4888-4341-b52d-b9b73013f9ad" Oct 03 16:40:46 crc kubenswrapper[4744]: E1003 16:40:46.368555 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2" podUID="d34995fe-607b-46ae-a2eb-c91d9c244415" Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.391868 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-psqc4"] Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.395969 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v"] Oct 03 16:40:46 crc kubenswrapper[4744]: E1003 16:40:46.415329 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:a82409e6d6a5554aad95acfe6fa4784e33de19a963eb8b1da1a80a3e6cf1ab55,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zc5wh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-555c7456bd-6xg8v_openstack-operators(438b2904-4b85-4d82-b883-f5788d4caab9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.421436 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw"] Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.433160 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89"] Oct 03 16:40:46 crc kubenswrapper[4744]: W1003 16:40:46.437667 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaef71e84_6035_42fb_a16f_c0e97876f1c0.slice/crio-75138268a9c79e631591a9e24a1ec735e3cdfd3acffd03868fdc540ceddbb2da WatchSource:0}: Error finding container 75138268a9c79e631591a9e24a1ec735e3cdfd3acffd03868fdc540ceddbb2da: Status 404 returned error can't find the container with id 75138268a9c79e631591a9e24a1ec735e3cdfd3acffd03868fdc540ceddbb2da Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.441225 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl"] Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.452745 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/911d5ffd-d447-4b8a-a177-7fc738d2aa35-cert\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678v879q\" (UID: \"911d5ffd-d447-4b8a-a177-7fc738d2aa35\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.453528 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-fcd7d9895-wjl4s"] Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.459586 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/911d5ffd-d447-4b8a-a177-7fc738d2aa35-cert\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d678v879q\" (UID: \"911d5ffd-d447-4b8a-a177-7fc738d2aa35\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" Oct 03 16:40:46 crc kubenswrapper[4744]: E1003 16:40:46.512587 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nhvnw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw_openstack-operators(aef71e84-6035-42fb-a16f-c0e97876f1c0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 16:40:46 crc kubenswrapper[4744]: E1003 16:40:46.514624 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw" podUID="aef71e84-6035-42fb-a16f-c0e97876f1c0" Oct 03 16:40:46 crc kubenswrapper[4744]: E1003 16:40:46.525359 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7tvfd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd5cb47d7-ksw89_openstack-operators(1d753e81-b04a-4216-991f-256a2ab6ac94): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 16:40:46 crc kubenswrapper[4744]: I1003 16:40:46.607296 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" Oct 03 16:40:46 crc kubenswrapper[4744]: E1003 16:40:46.892257 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v" podUID="438b2904-4b85-4d82-b883-f5788d4caab9" Oct 03 16:40:46 crc kubenswrapper[4744]: E1003 16:40:46.960585 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89" podUID="1d753e81-b04a-4216-991f-256a2ab6ac94" Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.073324 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" event={"ID":"3bc36b86-4888-4341-b52d-b9b73013f9ad","Type":"ContainerStarted","Data":"c3dd2500f79a8aa18c82a51fc1d7612385abeacc316a8eae008b4ca3601db696"} Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.074164 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" event={"ID":"3bc36b86-4888-4341-b52d-b9b73013f9ad","Type":"ContainerStarted","Data":"cd2996bd5a9c1aaa01fe543cecc19e23ab2ef86c86c95c5741af713670d1bfa8"} Oct 03 16:40:47 crc kubenswrapper[4744]: E1003 16:40:47.080069 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:40fb1819b6639807b77ef79448d35f1e4bfc1838a09d4f380e9fa0f755352475\\\"\"" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" podUID="3bc36b86-4888-4341-b52d-b9b73013f9ad" Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.080613 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-psqc4" event={"ID":"a6943fb4-ab7a-445f-b4d1-be9089562b77","Type":"ContainerStarted","Data":"775136e62f25ca4d5d598913a1c0283d83742d6c8e769c58f2bcec4e94ec9360"} Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.082430 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw" event={"ID":"aef71e84-6035-42fb-a16f-c0e97876f1c0","Type":"ContainerStarted","Data":"75138268a9c79e631591a9e24a1ec735e3cdfd3acffd03868fdc540ceddbb2da"} Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.085998 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-jktv9" event={"ID":"39de7155-617f-444a-bb16-0cb7fb229eb7","Type":"ContainerStarted","Data":"c1937ea9665a6039d0d0eb996964e234baf94c1b3621996b68d5cb27d8d72e17"} Oct 03 16:40:47 crc kubenswrapper[4744]: E1003 16:40:47.100420 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw" podUID="aef71e84-6035-42fb-a16f-c0e97876f1c0" Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.107453 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2" event={"ID":"d34995fe-607b-46ae-a2eb-c91d9c244415","Type":"ContainerStarted","Data":"e4cf616ebede78b80c8469760da9c3319c3ef19c678543176f70faa61e73a0aa"} Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.107488 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2" event={"ID":"d34995fe-607b-46ae-a2eb-c91d9c244415","Type":"ContainerStarted","Data":"67c58215ffacea8aba6828ce6ee2f8a50112236f5acb2d2bd177e393d0399a5e"} Oct 03 16:40:47 crc kubenswrapper[4744]: E1003 16:40:47.112635 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:110b885fe640ffdd8536e7da2a613677a6777e3d902e2ff15fa4d5968fe06c54\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2" podUID="d34995fe-607b-46ae-a2eb-c91d9c244415" Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.143688 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl" event={"ID":"4717fade-30fb-4f78-b7fb-dec2799d9b8c","Type":"ContainerStarted","Data":"5b8b21a57ca219fde55f60431fbbe30cc4e4abcc3113f1e77779c0c748e31b55"} Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.143757 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl" event={"ID":"4717fade-30fb-4f78-b7fb-dec2799d9b8c","Type":"ContainerStarted","Data":"4974171c3caa722c351f18a315080fbf93b834e6899a69f591ddd3be2927330f"} Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.149852 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-wjl4s" event={"ID":"1aa4d11e-aebd-477b-98e1-a05ea74a7677","Type":"ContainerStarted","Data":"f2a96f3fbb1559fe5d63114b6e005b5962f9ff9e35dab7ada266663b82d51e6b"} Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.151961 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm" event={"ID":"09c1f75f-d699-4675-b78d-a37353acd758","Type":"ContainerStarted","Data":"d6fabcc37b9f6c6854d79607ade9e3404b669f031c71a73b2500c12a3bb23cda"} Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.152007 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm" event={"ID":"09c1f75f-d699-4675-b78d-a37353acd758","Type":"ContainerStarted","Data":"13d291100112292d4ed3652a33af6181c4a07cb2f1fbc5992e1d76a59467bb08"} Oct 03 16:40:47 crc kubenswrapper[4744]: E1003 16:40:47.153687 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8f5eee2eb7b77432ef1a88ed693ff981514359dfc808581f393bcef252de5cfa\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm" podUID="09c1f75f-d699-4675-b78d-a37353acd758" Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.158886 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89" event={"ID":"1d753e81-b04a-4216-991f-256a2ab6ac94","Type":"ContainerStarted","Data":"6db5191c38b4544cfeab4065b864fa6c0f29dd041ed52023342450bc69292fa6"} Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.159049 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89" event={"ID":"1d753e81-b04a-4216-991f-256a2ab6ac94","Type":"ContainerStarted","Data":"2fa39450644dc45ec7f9ae4b9ad5df88c7241aedad0aeac7773a574bf4d25a46"} Oct 03 16:40:47 crc kubenswrapper[4744]: E1003 16:40:47.161348 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89" podUID="1d753e81-b04a-4216-991f-256a2ab6ac94" Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.163758 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v" event={"ID":"438b2904-4b85-4d82-b883-f5788d4caab9","Type":"ContainerStarted","Data":"e58e2486546a293b85dfdab9e1d01ef9675f26092284c4b19def6ec41e1d35cd"} Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.163835 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v" event={"ID":"438b2904-4b85-4d82-b883-f5788d4caab9","Type":"ContainerStarted","Data":"fb57044b29244cc3378bbdabb4f0e970ec57bc54b13cf1f620d0d457eafa06dd"} Oct 03 16:40:47 crc kubenswrapper[4744]: E1003 16:40:47.166713 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:a82409e6d6a5554aad95acfe6fa4784e33de19a963eb8b1da1a80a3e6cf1ab55\\\"\"" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v" podUID="438b2904-4b85-4d82-b883-f5788d4caab9" Oct 03 16:40:47 crc kubenswrapper[4744]: I1003 16:40:47.332774 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q"] Oct 03 16:40:48 crc kubenswrapper[4744]: I1003 16:40:48.194784 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" event={"ID":"911d5ffd-d447-4b8a-a177-7fc738d2aa35","Type":"ContainerStarted","Data":"c9608e2bbc5a10f0147d1defc2c70e840429c3d55e3b5b0abeda542a0f65a8df"} Oct 03 16:40:48 crc kubenswrapper[4744]: I1003 16:40:48.198524 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl" event={"ID":"4717fade-30fb-4f78-b7fb-dec2799d9b8c","Type":"ContainerStarted","Data":"38154f8e2f02dffccc65f9ac44cb53994479335436f554890d1e6cfb58e71610"} Oct 03 16:40:48 crc kubenswrapper[4744]: E1003 16:40:48.201191 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:110b885fe640ffdd8536e7da2a613677a6777e3d902e2ff15fa4d5968fe06c54\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2" podUID="d34995fe-607b-46ae-a2eb-c91d9c244415" Oct 03 16:40:48 crc kubenswrapper[4744]: E1003 16:40:48.201640 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89" podUID="1d753e81-b04a-4216-991f-256a2ab6ac94" Oct 03 16:40:48 crc kubenswrapper[4744]: E1003 16:40:48.201697 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:40fb1819b6639807b77ef79448d35f1e4bfc1838a09d4f380e9fa0f755352475\\\"\"" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" podUID="3bc36b86-4888-4341-b52d-b9b73013f9ad" Oct 03 16:40:48 crc kubenswrapper[4744]: E1003 16:40:48.201750 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8f5eee2eb7b77432ef1a88ed693ff981514359dfc808581f393bcef252de5cfa\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm" podUID="09c1f75f-d699-4675-b78d-a37353acd758" Oct 03 16:40:48 crc kubenswrapper[4744]: E1003 16:40:48.201789 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:a82409e6d6a5554aad95acfe6fa4784e33de19a963eb8b1da1a80a3e6cf1ab55\\\"\"" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v" podUID="438b2904-4b85-4d82-b883-f5788d4caab9" Oct 03 16:40:48 crc kubenswrapper[4744]: E1003 16:40:48.201832 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw" podUID="aef71e84-6035-42fb-a16f-c0e97876f1c0" Oct 03 16:40:48 crc kubenswrapper[4744]: I1003 16:40:48.281843 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl" podStartSLOduration=4.281816315 podStartE2EDuration="4.281816315s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:40:48.261048689 +0000 UTC m=+974.540924585" watchObservedRunningTime="2025-10-03 16:40:48.281816315 +0000 UTC m=+974.561692211" Oct 03 16:40:49 crc kubenswrapper[4744]: I1003 16:40:49.212197 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl" Oct 03 16:40:55 crc kubenswrapper[4744]: I1003 16:40:55.537232 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7758d4b55d-6xczl" Oct 03 16:40:59 crc kubenswrapper[4744]: E1003 16:40:59.201509 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:516f76ed86dd34225e6d0309451c7886bb81ff69032ba28125ae4d0cec54bce7" Oct 03 16:40:59 crc kubenswrapper[4744]: E1003 16:40:59.202329 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:516f76ed86dd34225e6d0309451c7886bb81ff69032ba28125ae4d0cec54bce7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-59p5n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7f55849f88-r8tdt_openstack-operators(a862283f-4243-42bd-bb2c-5b2392d657ac): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 16:40:59 crc kubenswrapper[4744]: E1003 16:40:59.400617 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt" podUID="a862283f-4243-42bd-bb2c-5b2392d657ac" Oct 03 16:41:00 crc kubenswrapper[4744]: I1003 16:41:00.293197 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-wjl4s" event={"ID":"1aa4d11e-aebd-477b-98e1-a05ea74a7677","Type":"ContainerStarted","Data":"a4caececd6a7c1d2576157c81ce4fc2bca3a5e666e4211c787d8e983ab08860c"} Oct 03 16:41:00 crc kubenswrapper[4744]: I1003 16:41:00.296253 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-xxdpr" event={"ID":"2832b841-b383-4eed-a8ce-71062269238a","Type":"ContainerStarted","Data":"8542ec1fd39222d727f4b09f4ec6d8ae7780ddeb319f6749fd23bb70f5e3eef6"} Oct 03 16:41:00 crc kubenswrapper[4744]: I1003 16:41:00.297586 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-jhqdv" event={"ID":"8be63e93-5f9d-4ffe-9ab2-d5b794c34f9b","Type":"ContainerStarted","Data":"3f9546e8fa29822f3b6c6153abcb803bf8069689f4b7ce9b753924366550f2bd"} Oct 03 16:41:00 crc kubenswrapper[4744]: I1003 16:41:00.299781 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" event={"ID":"911d5ffd-d447-4b8a-a177-7fc738d2aa35","Type":"ContainerStarted","Data":"b3e07678a8530ee9b2fe8930ead986ec06f5eeb24d15f97d7430800fe68d167d"} Oct 03 16:41:00 crc kubenswrapper[4744]: I1003 16:41:00.300847 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-4r792" event={"ID":"829ba77b-d3cc-416c-8813-eeff7b6b60bd","Type":"ContainerStarted","Data":"08f660bfed8348b18042b2fc395a46b7747a24905aa59aa71cf6288126080abd"} Oct 03 16:41:00 crc kubenswrapper[4744]: I1003 16:41:00.301977 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-n26jb" event={"ID":"fd8ecb24-30b8-4012-846a-4618f178996b","Type":"ContainerStarted","Data":"2a6aa3ea6e3676e0fb8a2bb1c7d5fd418fdfc725a8539935f8a2c0c94811643f"} Oct 03 16:41:00 crc kubenswrapper[4744]: I1003 16:41:00.303235 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-xr6db" event={"ID":"d34377f2-1b67-4c62-9602-ddb93f0cb4d7","Type":"ContainerStarted","Data":"498a1ae409b91afb6d343383a656d156fca49832a251532aa986e0638db42e52"} Oct 03 16:41:00 crc kubenswrapper[4744]: I1003 16:41:00.304190 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-bs7f6" event={"ID":"19371562-3c8d-4dce-864c-0a37cbce2bf3","Type":"ContainerStarted","Data":"1fe871d8b966648102597f8b8fdfce7d54ee3ede59f14eb6d645ef6e0a57d6f2"} Oct 03 16:41:00 crc kubenswrapper[4744]: I1003 16:41:00.305247 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-599898f689-fxkn9" event={"ID":"da0c0550-089f-4bd9-aaf8-960fcc75f019","Type":"ContainerStarted","Data":"44ea0a9e4f1302ecada374a8f09511816ebbed28ab49bf339d28f23c549cf3d3"} Oct 03 16:41:00 crc kubenswrapper[4744]: I1003 16:41:00.306660 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-4st8l" event={"ID":"d68c3557-6655-4449-8167-c28d1ccf442e","Type":"ContainerStarted","Data":"1ba910c44995676f06e4d6e1e5244f051f6c4ebc61dbbfe73903f2a26a9c3783"} Oct 03 16:41:00 crc kubenswrapper[4744]: I1003 16:41:00.308084 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-jktv9" event={"ID":"39de7155-617f-444a-bb16-0cb7fb229eb7","Type":"ContainerStarted","Data":"8bd3de86cf07466529d9f0a41c46cbefd1775f24daed9ebf187e437bfc6c537b"} Oct 03 16:41:00 crc kubenswrapper[4744]: I1003 16:41:00.309172 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt" event={"ID":"a862283f-4243-42bd-bb2c-5b2392d657ac","Type":"ContainerStarted","Data":"e609c07b668935fa4ec68bdac937c0196a6bd45af4b5f0012e2f70c13f0812e9"} Oct 03 16:41:00 crc kubenswrapper[4744]: E1003 16:41:00.310694 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:516f76ed86dd34225e6d0309451c7886bb81ff69032ba28125ae4d0cec54bce7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt" podUID="a862283f-4243-42bd-bb2c-5b2392d657ac" Oct 03 16:41:00 crc kubenswrapper[4744]: I1003 16:41:00.312253 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-tb5k2" event={"ID":"23519c20-6aec-4461-bea9-adaf88e87ab3","Type":"ContainerStarted","Data":"5ab054841ea998a071c63c242b2b4f2cfaac74f8694b45c83c8e1f2c821c6cc8"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.326053 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-jhqdv" event={"ID":"8be63e93-5f9d-4ffe-9ab2-d5b794c34f9b","Type":"ContainerStarted","Data":"d8b63bcd2c96a053dd91abc8bb964edcc98c1c7711147f02f99b0f13ed708882"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.328048 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-jhqdv" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.334129 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-xr6db" event={"ID":"d34377f2-1b67-4c62-9602-ddb93f0cb4d7","Type":"ContainerStarted","Data":"d3f9cb29c86223b29c11ae20035bcd6f9b5e4f8b6309c2c76297caed2e6d6cc2"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.334931 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-xr6db" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.337673 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-wjl4s" event={"ID":"1aa4d11e-aebd-477b-98e1-a05ea74a7677","Type":"ContainerStarted","Data":"632085db213f9444f27d70bc43d5038a02678d135b9e48aa3c4e06879cc487f2"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.338101 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-wjl4s" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.346782 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-xxdpr" event={"ID":"2832b841-b383-4eed-a8ce-71062269238a","Type":"ContainerStarted","Data":"4c4c036e4334f7549daa18303e71168ba442069518ab1fcde42da8519c4b41fd"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.348025 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-xxdpr" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.350896 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-zd5fc" event={"ID":"be9443a6-46c8-4374-8caa-8040ae117981","Type":"ContainerStarted","Data":"3f3d95ec2efec1b86ceea455c37ec7f5020113cdca9f3ff5b44c24f0955f18e8"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.350926 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-zd5fc" event={"ID":"be9443a6-46c8-4374-8caa-8040ae117981","Type":"ContainerStarted","Data":"68043314abebf0e990c8c8d5475816036dfa13ee89a05cda4562e1d8fa1ee069"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.351376 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-zd5fc" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.353985 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-tb5k2" event={"ID":"23519c20-6aec-4461-bea9-adaf88e87ab3","Type":"ContainerStarted","Data":"e68e235c38bd04a89bf3a1c31461fcd2b3aa9e80856ff912cea7b3cebd54ac70"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.354235 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-tb5k2" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.366244 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" event={"ID":"911d5ffd-d447-4b8a-a177-7fc738d2aa35","Type":"ContainerStarted","Data":"63d5b06a0007dfb563af5971a42179a9f31788063ef846db145fee34f95f50ab"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.367302 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.367818 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-jhqdv" podStartSLOduration=4.0315749 podStartE2EDuration="17.36777933s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:45.964778905 +0000 UTC m=+972.244654801" lastFinishedPulling="2025-10-03 16:40:59.300983335 +0000 UTC m=+985.580859231" observedRunningTime="2025-10-03 16:41:01.358690306 +0000 UTC m=+987.638566212" watchObservedRunningTime="2025-10-03 16:41:01.36777933 +0000 UTC m=+987.647655236" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.373788 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-n26jb" event={"ID":"fd8ecb24-30b8-4012-846a-4618f178996b","Type":"ContainerStarted","Data":"e16f0b01428e9b937c16f45d2bc96f903b485c9ea738165a75a962394666dc47"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.373854 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-n26jb" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.377150 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-bs7f6" event={"ID":"19371562-3c8d-4dce-864c-0a37cbce2bf3","Type":"ContainerStarted","Data":"47154b7df169000f47d96f4f969ee756c329044f351dc631665a7e9fc4955471"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.378103 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-bs7f6" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.386779 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-psqc4" event={"ID":"a6943fb4-ab7a-445f-b4d1-be9089562b77","Type":"ContainerStarted","Data":"edc1689b8a91b606f366c653f2de3f39597f4f47377ae698c65c9145b8f00a9e"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.386914 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-psqc4" event={"ID":"a6943fb4-ab7a-445f-b4d1-be9089562b77","Type":"ContainerStarted","Data":"e172e959b721db1cba743ab06ab2eb026c60ca6b283afc9f71f8a73d54186054"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.387009 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-psqc4" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.397419 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-xr6db" podStartSLOduration=4.051438572 podStartE2EDuration="17.397385465s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:45.977127592 +0000 UTC m=+972.257003498" lastFinishedPulling="2025-10-03 16:40:59.323074495 +0000 UTC m=+985.602950391" observedRunningTime="2025-10-03 16:41:01.390740703 +0000 UTC m=+987.670616609" watchObservedRunningTime="2025-10-03 16:41:01.397385465 +0000 UTC m=+987.677261361" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.406528 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-4st8l" event={"ID":"d68c3557-6655-4449-8167-c28d1ccf442e","Type":"ContainerStarted","Data":"4bc80ff59f551d45dd76e04be7d0d31ee543dc00e051ca1938827d1229e99d68"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.408012 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-4st8l" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.424018 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-4r792" event={"ID":"829ba77b-d3cc-416c-8813-eeff7b6b60bd","Type":"ContainerStarted","Data":"7b00204c86e77318180a456ae119c16007ab7fc462d6a0b6b501ec44cedd9345"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.425734 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-wjl4s" podStartSLOduration=4.629731696 podStartE2EDuration="17.425709396s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:46.499567684 +0000 UTC m=+972.779443580" lastFinishedPulling="2025-10-03 16:40:59.295545384 +0000 UTC m=+985.575421280" observedRunningTime="2025-10-03 16:41:01.424765702 +0000 UTC m=+987.704641598" watchObservedRunningTime="2025-10-03 16:41:01.425709396 +0000 UTC m=+987.705585292" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.426802 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-4r792" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.429852 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-599898f689-fxkn9" event={"ID":"da0c0550-089f-4bd9-aaf8-960fcc75f019","Type":"ContainerStarted","Data":"2908b7d3fed2faf46b290afcc05791e85febd7965f3fd01132f3a99189090021"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.430449 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-599898f689-fxkn9" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.460290 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-7qtkn" event={"ID":"a38be30f-e956-4859-9610-a4d0735fe9d4","Type":"ContainerStarted","Data":"e61e16b7dce5f9d84e199098ebae143a613f98dc17081ffbf69438e8000a8139"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.460374 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-7qtkn" event={"ID":"a38be30f-e956-4859-9610-a4d0735fe9d4","Type":"ContainerStarted","Data":"db3dc85f5807611fa84dcdb0be16aefe5ccb2b4f2477481716381e69c81a07cd"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.460818 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-7qtkn" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.460999 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-zd5fc" podStartSLOduration=4.127678431 podStartE2EDuration="17.460973526s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:45.96962787 +0000 UTC m=+972.249503766" lastFinishedPulling="2025-10-03 16:40:59.302922955 +0000 UTC m=+985.582798861" observedRunningTime="2025-10-03 16:41:01.453511964 +0000 UTC m=+987.733387870" watchObservedRunningTime="2025-10-03 16:41:01.460973526 +0000 UTC m=+987.740849422" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.476037 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-jktv9" event={"ID":"39de7155-617f-444a-bb16-0cb7fb229eb7","Type":"ContainerStarted","Data":"f0599da4059f04896ab3b38258f1b57aed53a1fffe5031c27f84633c0ffc0dc4"} Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.476082 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-jktv9" Oct 03 16:41:01 crc kubenswrapper[4744]: E1003 16:41:01.484896 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:516f76ed86dd34225e6d0309451c7886bb81ff69032ba28125ae4d0cec54bce7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt" podUID="a862283f-4243-42bd-bb2c-5b2392d657ac" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.490995 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-tb5k2" podStartSLOduration=4.097581493 podStartE2EDuration="17.490961111s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:45.909737012 +0000 UTC m=+972.189612908" lastFinishedPulling="2025-10-03 16:40:59.30311663 +0000 UTC m=+985.582992526" observedRunningTime="2025-10-03 16:41:01.483846937 +0000 UTC m=+987.763722833" watchObservedRunningTime="2025-10-03 16:41:01.490961111 +0000 UTC m=+987.770837007" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.535681 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-xxdpr" podStartSLOduration=4.208353154 podStartE2EDuration="17.535658295s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:45.971344864 +0000 UTC m=+972.251220760" lastFinishedPulling="2025-10-03 16:40:59.298650004 +0000 UTC m=+985.578525901" observedRunningTime="2025-10-03 16:41:01.530305546 +0000 UTC m=+987.810181442" watchObservedRunningTime="2025-10-03 16:41:01.535658295 +0000 UTC m=+987.815534191" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.581839 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-psqc4" podStartSLOduration=4.688389653 podStartE2EDuration="17.581826456s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:46.407013948 +0000 UTC m=+972.686889844" lastFinishedPulling="2025-10-03 16:40:59.300450751 +0000 UTC m=+985.580326647" observedRunningTime="2025-10-03 16:41:01.581351914 +0000 UTC m=+987.861227810" watchObservedRunningTime="2025-10-03 16:41:01.581826456 +0000 UTC m=+987.861702352" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.601648 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-n26jb" podStartSLOduration=4.290994438 podStartE2EDuration="17.601631898s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:45.975613344 +0000 UTC m=+972.255489240" lastFinishedPulling="2025-10-03 16:40:59.286250784 +0000 UTC m=+985.566126700" observedRunningTime="2025-10-03 16:41:01.600826627 +0000 UTC m=+987.880702523" watchObservedRunningTime="2025-10-03 16:41:01.601631898 +0000 UTC m=+987.881507794" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.621271 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-jktv9" podStartSLOduration=4.444234923 podStartE2EDuration="17.621248914s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:46.109281235 +0000 UTC m=+972.389157121" lastFinishedPulling="2025-10-03 16:40:59.286295206 +0000 UTC m=+985.566171112" observedRunningTime="2025-10-03 16:41:01.616936253 +0000 UTC m=+987.896812149" watchObservedRunningTime="2025-10-03 16:41:01.621248914 +0000 UTC m=+987.901124800" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.643099 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-7qtkn" podStartSLOduration=4.25083783 podStartE2EDuration="17.643063857s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:45.909821455 +0000 UTC m=+972.189697351" lastFinishedPulling="2025-10-03 16:40:59.302047492 +0000 UTC m=+985.581923378" observedRunningTime="2025-10-03 16:41:01.640240134 +0000 UTC m=+987.920116040" watchObservedRunningTime="2025-10-03 16:41:01.643063857 +0000 UTC m=+987.922939753" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.675383 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-bs7f6" podStartSLOduration=4.256938131 podStartE2EDuration="17.675359901s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:45.885594953 +0000 UTC m=+972.165470839" lastFinishedPulling="2025-10-03 16:40:59.304016713 +0000 UTC m=+985.583892609" observedRunningTime="2025-10-03 16:41:01.66293974 +0000 UTC m=+987.942815636" watchObservedRunningTime="2025-10-03 16:41:01.675359901 +0000 UTC m=+987.955235797" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.698255 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" podStartSLOduration=5.734013125 podStartE2EDuration="17.698240512s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:47.336901051 +0000 UTC m=+973.616776947" lastFinishedPulling="2025-10-03 16:40:59.301128438 +0000 UTC m=+985.581004334" observedRunningTime="2025-10-03 16:41:01.69311957 +0000 UTC m=+987.972995466" watchObservedRunningTime="2025-10-03 16:41:01.698240512 +0000 UTC m=+987.978116408" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.712932 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-599898f689-fxkn9" podStartSLOduration=4.363488281 podStartE2EDuration="17.71290918s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:45.954789259 +0000 UTC m=+972.234665155" lastFinishedPulling="2025-10-03 16:40:59.304210158 +0000 UTC m=+985.584086054" observedRunningTime="2025-10-03 16:41:01.710424826 +0000 UTC m=+987.990300722" watchObservedRunningTime="2025-10-03 16:41:01.71290918 +0000 UTC m=+987.992785076" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.736141 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-4st8l" podStartSLOduration=3.8107837890000003 podStartE2EDuration="17.73612229s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:45.300120854 +0000 UTC m=+971.579996750" lastFinishedPulling="2025-10-03 16:40:59.225459355 +0000 UTC m=+985.505335251" observedRunningTime="2025-10-03 16:41:01.73575574 +0000 UTC m=+988.015631636" watchObservedRunningTime="2025-10-03 16:41:01.73612229 +0000 UTC m=+988.015998206" Oct 03 16:41:01 crc kubenswrapper[4744]: I1003 16:41:01.769207 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-4r792" podStartSLOduration=4.1475198970000005 podStartE2EDuration="17.769191313s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:45.603710757 +0000 UTC m=+971.883586653" lastFinishedPulling="2025-10-03 16:40:59.225382163 +0000 UTC m=+985.505258069" observedRunningTime="2025-10-03 16:41:01.764642696 +0000 UTC m=+988.044518592" watchObservedRunningTime="2025-10-03 16:41:01.769191313 +0000 UTC m=+988.049067209" Oct 03 16:41:04 crc kubenswrapper[4744]: I1003 16:41:04.478176 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-4st8l" Oct 03 16:41:04 crc kubenswrapper[4744]: I1003 16:41:04.496941 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-4r792" Oct 03 16:41:04 crc kubenswrapper[4744]: I1003 16:41:04.538125 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-bs7f6" Oct 03 16:41:04 crc kubenswrapper[4744]: I1003 16:41:04.582075 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-599898f689-fxkn9" Oct 03 16:41:04 crc kubenswrapper[4744]: I1003 16:41:04.668334 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:41:04 crc kubenswrapper[4744]: I1003 16:41:04.668411 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:41:04 crc kubenswrapper[4744]: I1003 16:41:04.668817 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:41:04 crc kubenswrapper[4744]: I1003 16:41:04.669830 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7ad0acf841a19f4f4edda47a15b6310ad0b8ad88f8fb721a55d7cdc8e8814147"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 16:41:04 crc kubenswrapper[4744]: I1003 16:41:04.669905 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://7ad0acf841a19f4f4edda47a15b6310ad0b8ad88f8fb721a55d7cdc8e8814147" gracePeriod=600 Oct 03 16:41:04 crc kubenswrapper[4744]: I1003 16:41:04.902890 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-n26jb" Oct 03 16:41:04 crc kubenswrapper[4744]: I1003 16:41:04.911041 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-xxdpr" Oct 03 16:41:05 crc kubenswrapper[4744]: I1003 16:41:05.100569 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-xr6db" Oct 03 16:41:05 crc kubenswrapper[4744]: I1003 16:41:05.156250 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-jhqdv" Oct 03 16:41:05 crc kubenswrapper[4744]: I1003 16:41:05.173616 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-jktv9" Oct 03 16:41:05 crc kubenswrapper[4744]: I1003 16:41:05.184397 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-psqc4" Oct 03 16:41:05 crc kubenswrapper[4744]: I1003 16:41:05.518090 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-wjl4s" Oct 03 16:41:05 crc kubenswrapper[4744]: I1003 16:41:05.520530 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="7ad0acf841a19f4f4edda47a15b6310ad0b8ad88f8fb721a55d7cdc8e8814147" exitCode=0 Oct 03 16:41:05 crc kubenswrapper[4744]: I1003 16:41:05.520561 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"7ad0acf841a19f4f4edda47a15b6310ad0b8ad88f8fb721a55d7cdc8e8814147"} Oct 03 16:41:05 crc kubenswrapper[4744]: I1003 16:41:05.520587 4744 scope.go:117] "RemoveContainer" containerID="08b50c5cd69a2f26327ea9dc4a127dffa53b9765528a51178793482f7f49b17f" Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.530714 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v" event={"ID":"438b2904-4b85-4d82-b883-f5788d4caab9","Type":"ContainerStarted","Data":"9a14d50137363c7c0e508aeb464dded5d4c378db45fe4643ffa98a9808087c20"} Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.532438 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw" event={"ID":"aef71e84-6035-42fb-a16f-c0e97876f1c0","Type":"ContainerStarted","Data":"f10e0962acf15b836a87f2463cfd2fad73d8a5ade163252f164ae2a45072cf79"} Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.532678 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v" Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.534290 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" event={"ID":"3bc36b86-4888-4341-b52d-b9b73013f9ad","Type":"ContainerStarted","Data":"071937a9908de5152b055f3b60701e614b7c6b79ee56c3589ebdcdfc2bdb1710"} Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.534536 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.536428 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2" event={"ID":"d34995fe-607b-46ae-a2eb-c91d9c244415","Type":"ContainerStarted","Data":"1f8dae8dd5085ccb47981ab8e4c2532711aa3e9bd64675f15048dd9ef339dc44"} Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.536650 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2" Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.539446 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"7b3e699f99d905cb6203d07f6a9bf0a0b414b4e9270f595c5bdaf98f6b96e23b"} Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.541332 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm" event={"ID":"09c1f75f-d699-4675-b78d-a37353acd758","Type":"ContainerStarted","Data":"50ecd5eb37e02a39f579bcd157edbe05424bb33e7943763b3fe5c2eddd105c9f"} Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.541527 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm" Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.542995 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89" event={"ID":"1d753e81-b04a-4216-991f-256a2ab6ac94","Type":"ContainerStarted","Data":"456a1f6c30ec46116897a340c8edaac599af7d9a16f5c901f9d5424bc05a1b0b"} Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.543186 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89" Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.589346 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89" podStartSLOduration=3.658567841 podStartE2EDuration="22.58932213s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:46.52512595 +0000 UTC m=+972.805001846" lastFinishedPulling="2025-10-03 16:41:05.455880239 +0000 UTC m=+991.735756135" observedRunningTime="2025-10-03 16:41:06.5850828 +0000 UTC m=+992.864958736" watchObservedRunningTime="2025-10-03 16:41:06.58932213 +0000 UTC m=+992.869198046" Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.596483 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v" podStartSLOduration=3.571888719 podStartE2EDuration="22.596463664s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:46.415123086 +0000 UTC m=+972.694998982" lastFinishedPulling="2025-10-03 16:41:05.439698031 +0000 UTC m=+991.719573927" observedRunningTime="2025-10-03 16:41:06.559162531 +0000 UTC m=+992.839038427" watchObservedRunningTime="2025-10-03 16:41:06.596463664 +0000 UTC m=+992.876339570" Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.616801 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d678v879q" Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.633825 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" podStartSLOduration=3.331962146 podStartE2EDuration="22.633804348s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:46.136166815 +0000 UTC m=+972.416042711" lastFinishedPulling="2025-10-03 16:41:05.438009017 +0000 UTC m=+991.717884913" observedRunningTime="2025-10-03 16:41:06.631891049 +0000 UTC m=+992.911766965" watchObservedRunningTime="2025-10-03 16:41:06.633804348 +0000 UTC m=+992.913680244" Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.654026 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2" podStartSLOduration=3.355658435 podStartE2EDuration="22.65400565s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:46.156508117 +0000 UTC m=+972.436384013" lastFinishedPulling="2025-10-03 16:41:05.454855332 +0000 UTC m=+991.734731228" observedRunningTime="2025-10-03 16:41:06.650007956 +0000 UTC m=+992.929883852" watchObservedRunningTime="2025-10-03 16:41:06.65400565 +0000 UTC m=+992.933881556" Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.671603 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm" podStartSLOduration=3.361537738 podStartE2EDuration="22.671587543s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:46.146596153 +0000 UTC m=+972.426472049" lastFinishedPulling="2025-10-03 16:41:05.456645958 +0000 UTC m=+991.736521854" observedRunningTime="2025-10-03 16:41:06.665221969 +0000 UTC m=+992.945097875" watchObservedRunningTime="2025-10-03 16:41:06.671587543 +0000 UTC m=+992.951463429" Oct 03 16:41:06 crc kubenswrapper[4744]: I1003 16:41:06.684965 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw" podStartSLOduration=3.741405001 podStartE2EDuration="22.684947418s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:46.51228583 +0000 UTC m=+972.792161726" lastFinishedPulling="2025-10-03 16:41:05.455828247 +0000 UTC m=+991.735704143" observedRunningTime="2025-10-03 16:41:06.684083636 +0000 UTC m=+992.963959552" watchObservedRunningTime="2025-10-03 16:41:06.684947418 +0000 UTC m=+992.964823324" Oct 03 16:41:14 crc kubenswrapper[4744]: I1003 16:41:14.541545 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-7qtkn" Oct 03 16:41:14 crc kubenswrapper[4744]: I1003 16:41:14.651371 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-tb5k2" Oct 03 16:41:14 crc kubenswrapper[4744]: I1003 16:41:14.851667 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-zd5fc" Oct 03 16:41:14 crc kubenswrapper[4744]: I1003 16:41:14.886930 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-kzsg2" Oct 03 16:41:15 crc kubenswrapper[4744]: I1003 16:41:15.197435 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-7qgsm" Oct 03 16:41:15 crc kubenswrapper[4744]: I1003 16:41:15.224339 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-4jdmt" Oct 03 16:41:15 crc kubenswrapper[4744]: I1003 16:41:15.226826 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-ksw89" Oct 03 16:41:15 crc kubenswrapper[4744]: I1003 16:41:15.282448 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-6xg8v" Oct 03 16:41:15 crc kubenswrapper[4744]: I1003 16:41:15.622838 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt" event={"ID":"a862283f-4243-42bd-bb2c-5b2392d657ac","Type":"ContainerStarted","Data":"58b6d8bb1e86e96d41d3521e60b42043de1866f2948470121078c3db283ad9c1"} Oct 03 16:41:15 crc kubenswrapper[4744]: I1003 16:41:15.623658 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt" Oct 03 16:41:15 crc kubenswrapper[4744]: I1003 16:41:15.642802 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt" podStartSLOduration=2.194346802 podStartE2EDuration="31.642776762s" podCreationTimestamp="2025-10-03 16:40:44 +0000 UTC" firstStartedPulling="2025-10-03 16:40:45.900361732 +0000 UTC m=+972.180237628" lastFinishedPulling="2025-10-03 16:41:15.348791692 +0000 UTC m=+1001.628667588" observedRunningTime="2025-10-03 16:41:15.640315628 +0000 UTC m=+1001.920191524" watchObservedRunningTime="2025-10-03 16:41:15.642776762 +0000 UTC m=+1001.922652678" Oct 03 16:41:24 crc kubenswrapper[4744]: I1003 16:41:24.729404 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-r8tdt" Oct 03 16:41:40 crc kubenswrapper[4744]: I1003 16:41:40.991922 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-hvq55"] Oct 03 16:41:40 crc kubenswrapper[4744]: I1003 16:41:40.993529 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-hvq55" Oct 03 16:41:40 crc kubenswrapper[4744]: I1003 16:41:40.996775 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-8gksm" Oct 03 16:41:40 crc kubenswrapper[4744]: I1003 16:41:40.997137 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.012634 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-hvq55"] Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.109143 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6ksgr"] Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.113662 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.118047 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6ksgr"] Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.120099 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvhgq\" (UniqueName: \"kubernetes.io/projected/9d8609ec-6619-4d44-a1f2-3b91146766ba-kube-api-access-fvhgq\") pod \"dnsmasq-dns-675f4bcbfc-hvq55\" (UID: \"9d8609ec-6619-4d44-a1f2-3b91146766ba\") " pod="openstack/dnsmasq-dns-675f4bcbfc-hvq55" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.120154 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8609ec-6619-4d44-a1f2-3b91146766ba-config\") pod \"dnsmasq-dns-675f4bcbfc-hvq55\" (UID: \"9d8609ec-6619-4d44-a1f2-3b91146766ba\") " pod="openstack/dnsmasq-dns-675f4bcbfc-hvq55" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.120439 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.222193 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8609ec-6619-4d44-a1f2-3b91146766ba-config\") pod \"dnsmasq-dns-675f4bcbfc-hvq55\" (UID: \"9d8609ec-6619-4d44-a1f2-3b91146766ba\") " pod="openstack/dnsmasq-dns-675f4bcbfc-hvq55" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.222305 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e8436a4-6ba0-4112-84ea-4c40491cea31-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-6ksgr\" (UID: \"7e8436a4-6ba0-4112-84ea-4c40491cea31\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.222328 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsgnp\" (UniqueName: \"kubernetes.io/projected/7e8436a4-6ba0-4112-84ea-4c40491cea31-kube-api-access-gsgnp\") pod \"dnsmasq-dns-78dd6ddcc-6ksgr\" (UID: \"7e8436a4-6ba0-4112-84ea-4c40491cea31\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.222368 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e8436a4-6ba0-4112-84ea-4c40491cea31-config\") pod \"dnsmasq-dns-78dd6ddcc-6ksgr\" (UID: \"7e8436a4-6ba0-4112-84ea-4c40491cea31\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.222455 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvhgq\" (UniqueName: \"kubernetes.io/projected/9d8609ec-6619-4d44-a1f2-3b91146766ba-kube-api-access-fvhgq\") pod \"dnsmasq-dns-675f4bcbfc-hvq55\" (UID: \"9d8609ec-6619-4d44-a1f2-3b91146766ba\") " pod="openstack/dnsmasq-dns-675f4bcbfc-hvq55" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.224256 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8609ec-6619-4d44-a1f2-3b91146766ba-config\") pod \"dnsmasq-dns-675f4bcbfc-hvq55\" (UID: \"9d8609ec-6619-4d44-a1f2-3b91146766ba\") " pod="openstack/dnsmasq-dns-675f4bcbfc-hvq55" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.254779 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvhgq\" (UniqueName: \"kubernetes.io/projected/9d8609ec-6619-4d44-a1f2-3b91146766ba-kube-api-access-fvhgq\") pod \"dnsmasq-dns-675f4bcbfc-hvq55\" (UID: \"9d8609ec-6619-4d44-a1f2-3b91146766ba\") " pod="openstack/dnsmasq-dns-675f4bcbfc-hvq55" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.312232 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-hvq55" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.324352 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsgnp\" (UniqueName: \"kubernetes.io/projected/7e8436a4-6ba0-4112-84ea-4c40491cea31-kube-api-access-gsgnp\") pod \"dnsmasq-dns-78dd6ddcc-6ksgr\" (UID: \"7e8436a4-6ba0-4112-84ea-4c40491cea31\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.324401 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e8436a4-6ba0-4112-84ea-4c40491cea31-config\") pod \"dnsmasq-dns-78dd6ddcc-6ksgr\" (UID: \"7e8436a4-6ba0-4112-84ea-4c40491cea31\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.324539 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e8436a4-6ba0-4112-84ea-4c40491cea31-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-6ksgr\" (UID: \"7e8436a4-6ba0-4112-84ea-4c40491cea31\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.325308 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e8436a4-6ba0-4112-84ea-4c40491cea31-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-6ksgr\" (UID: \"7e8436a4-6ba0-4112-84ea-4c40491cea31\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.325450 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e8436a4-6ba0-4112-84ea-4c40491cea31-config\") pod \"dnsmasq-dns-78dd6ddcc-6ksgr\" (UID: \"7e8436a4-6ba0-4112-84ea-4c40491cea31\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.343650 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsgnp\" (UniqueName: \"kubernetes.io/projected/7e8436a4-6ba0-4112-84ea-4c40491cea31-kube-api-access-gsgnp\") pod \"dnsmasq-dns-78dd6ddcc-6ksgr\" (UID: \"7e8436a4-6ba0-4112-84ea-4c40491cea31\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.431700 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.773171 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-hvq55"] Oct 03 16:41:41 crc kubenswrapper[4744]: W1003 16:41:41.781443 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d8609ec_6619_4d44_a1f2_3b91146766ba.slice/crio-aa517e0ffbd19e44ee8b04aed7e03721b8f5eb9693aeeefcec699703908fb0e4 WatchSource:0}: Error finding container aa517e0ffbd19e44ee8b04aed7e03721b8f5eb9693aeeefcec699703908fb0e4: Status 404 returned error can't find the container with id aa517e0ffbd19e44ee8b04aed7e03721b8f5eb9693aeeefcec699703908fb0e4 Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.855795 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-hvq55" event={"ID":"9d8609ec-6619-4d44-a1f2-3b91146766ba","Type":"ContainerStarted","Data":"aa517e0ffbd19e44ee8b04aed7e03721b8f5eb9693aeeefcec699703908fb0e4"} Oct 03 16:41:41 crc kubenswrapper[4744]: I1003 16:41:41.902964 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6ksgr"] Oct 03 16:41:41 crc kubenswrapper[4744]: W1003 16:41:41.906189 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e8436a4_6ba0_4112_84ea_4c40491cea31.slice/crio-fcffff7dd791184eff9d72e347ed1ec14a2c985e5c1950d45a519dbabd66630f WatchSource:0}: Error finding container fcffff7dd791184eff9d72e347ed1ec14a2c985e5c1950d45a519dbabd66630f: Status 404 returned error can't find the container with id fcffff7dd791184eff9d72e347ed1ec14a2c985e5c1950d45a519dbabd66630f Oct 03 16:41:42 crc kubenswrapper[4744]: I1003 16:41:42.863086 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" event={"ID":"7e8436a4-6ba0-4112-84ea-4c40491cea31","Type":"ContainerStarted","Data":"fcffff7dd791184eff9d72e347ed1ec14a2c985e5c1950d45a519dbabd66630f"} Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.096078 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-hvq55"] Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.124073 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-64zgj"] Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.129395 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-64zgj" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.162637 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-64zgj"] Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.279323 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg28q\" (UniqueName: \"kubernetes.io/projected/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-kube-api-access-wg28q\") pod \"dnsmasq-dns-666b6646f7-64zgj\" (UID: \"ccc57209-f9dc-4308-b11e-d3e10c70fe0a\") " pod="openstack/dnsmasq-dns-666b6646f7-64zgj" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.279368 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-dns-svc\") pod \"dnsmasq-dns-666b6646f7-64zgj\" (UID: \"ccc57209-f9dc-4308-b11e-d3e10c70fe0a\") " pod="openstack/dnsmasq-dns-666b6646f7-64zgj" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.279549 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-config\") pod \"dnsmasq-dns-666b6646f7-64zgj\" (UID: \"ccc57209-f9dc-4308-b11e-d3e10c70fe0a\") " pod="openstack/dnsmasq-dns-666b6646f7-64zgj" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.380595 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-config\") pod \"dnsmasq-dns-666b6646f7-64zgj\" (UID: \"ccc57209-f9dc-4308-b11e-d3e10c70fe0a\") " pod="openstack/dnsmasq-dns-666b6646f7-64zgj" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.380689 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg28q\" (UniqueName: \"kubernetes.io/projected/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-kube-api-access-wg28q\") pod \"dnsmasq-dns-666b6646f7-64zgj\" (UID: \"ccc57209-f9dc-4308-b11e-d3e10c70fe0a\") " pod="openstack/dnsmasq-dns-666b6646f7-64zgj" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.380707 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-dns-svc\") pod \"dnsmasq-dns-666b6646f7-64zgj\" (UID: \"ccc57209-f9dc-4308-b11e-d3e10c70fe0a\") " pod="openstack/dnsmasq-dns-666b6646f7-64zgj" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.381754 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-dns-svc\") pod \"dnsmasq-dns-666b6646f7-64zgj\" (UID: \"ccc57209-f9dc-4308-b11e-d3e10c70fe0a\") " pod="openstack/dnsmasq-dns-666b6646f7-64zgj" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.382276 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-config\") pod \"dnsmasq-dns-666b6646f7-64zgj\" (UID: \"ccc57209-f9dc-4308-b11e-d3e10c70fe0a\") " pod="openstack/dnsmasq-dns-666b6646f7-64zgj" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.410642 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6ksgr"] Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.437765 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg28q\" (UniqueName: \"kubernetes.io/projected/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-kube-api-access-wg28q\") pod \"dnsmasq-dns-666b6646f7-64zgj\" (UID: \"ccc57209-f9dc-4308-b11e-d3e10c70fe0a\") " pod="openstack/dnsmasq-dns-666b6646f7-64zgj" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.462993 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bwpkl"] Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.464460 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.470844 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bwpkl"] Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.487824 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-64zgj" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.587442 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19d490cc-8f49-4957-9812-94e96b24e100-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bwpkl\" (UID: \"19d490cc-8f49-4957-9812-94e96b24e100\") " pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.587551 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtn2l\" (UniqueName: \"kubernetes.io/projected/19d490cc-8f49-4957-9812-94e96b24e100-kube-api-access-qtn2l\") pod \"dnsmasq-dns-57d769cc4f-bwpkl\" (UID: \"19d490cc-8f49-4957-9812-94e96b24e100\") " pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.587580 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d490cc-8f49-4957-9812-94e96b24e100-config\") pod \"dnsmasq-dns-57d769cc4f-bwpkl\" (UID: \"19d490cc-8f49-4957-9812-94e96b24e100\") " pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.688891 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19d490cc-8f49-4957-9812-94e96b24e100-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bwpkl\" (UID: \"19d490cc-8f49-4957-9812-94e96b24e100\") " pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.688958 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtn2l\" (UniqueName: \"kubernetes.io/projected/19d490cc-8f49-4957-9812-94e96b24e100-kube-api-access-qtn2l\") pod \"dnsmasq-dns-57d769cc4f-bwpkl\" (UID: \"19d490cc-8f49-4957-9812-94e96b24e100\") " pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.688982 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d490cc-8f49-4957-9812-94e96b24e100-config\") pod \"dnsmasq-dns-57d769cc4f-bwpkl\" (UID: \"19d490cc-8f49-4957-9812-94e96b24e100\") " pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.689753 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19d490cc-8f49-4957-9812-94e96b24e100-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bwpkl\" (UID: \"19d490cc-8f49-4957-9812-94e96b24e100\") " pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.689993 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d490cc-8f49-4957-9812-94e96b24e100-config\") pod \"dnsmasq-dns-57d769cc4f-bwpkl\" (UID: \"19d490cc-8f49-4957-9812-94e96b24e100\") " pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.705379 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtn2l\" (UniqueName: \"kubernetes.io/projected/19d490cc-8f49-4957-9812-94e96b24e100-kube-api-access-qtn2l\") pod \"dnsmasq-dns-57d769cc4f-bwpkl\" (UID: \"19d490cc-8f49-4957-9812-94e96b24e100\") " pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" Oct 03 16:41:44 crc kubenswrapper[4744]: I1003 16:41:44.800314 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.290281 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.292216 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.294973 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.295004 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.295068 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.295143 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.295237 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.295308 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.295789 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-wjds9" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.303737 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.398938 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d21c39f-eece-4cf0-847b-c8ebc87c574d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.399197 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.399313 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d21c39f-eece-4cf0-847b-c8ebc87c574d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.399387 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.399484 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.399607 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-config-data\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.399707 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.399783 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv2nf\" (UniqueName: \"kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-kube-api-access-cv2nf\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.399849 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.399935 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.400011 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.501805 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.501866 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-config-data\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.501947 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.501973 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv2nf\" (UniqueName: \"kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-kube-api-access-cv2nf\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.501993 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.502017 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.502048 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.502146 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d21c39f-eece-4cf0-847b-c8ebc87c574d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.502180 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.502214 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d21c39f-eece-4cf0-847b-c8ebc87c574d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.502244 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.502466 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.502856 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.502904 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.503423 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-config-data\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.503791 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.504114 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.507620 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d21c39f-eece-4cf0-847b-c8ebc87c574d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.507951 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.508792 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.508908 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d21c39f-eece-4cf0-847b-c8ebc87c574d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.521798 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv2nf\" (UniqueName: \"kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-kube-api-access-cv2nf\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.535425 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.579610 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.582372 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.585040 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.585086 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.585171 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-7d2qj" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.585364 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.585391 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.585821 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.587645 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.592129 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.648254 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.705396 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.705473 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.705547 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.705618 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.705649 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.705690 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.705720 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hphw\" (UniqueName: \"kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-kube-api-access-6hphw\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.705759 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.705789 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.705810 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.705957 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.807397 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.807449 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.807488 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.807528 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.807546 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.807573 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hphw\" (UniqueName: \"kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-kube-api-access-6hphw\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.807591 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.807619 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.807635 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.807653 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.807682 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.808187 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.809395 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.809882 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.810032 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.810800 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.811560 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.830841 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.830994 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.831882 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hphw\" (UniqueName: \"kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-kube-api-access-6hphw\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.835359 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.837002 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.845757 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:45 crc kubenswrapper[4744]: I1003 16:41:45.909046 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.720785 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.723481 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.727247 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.727923 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.727988 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-9zq7q" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.728130 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.728281 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.739157 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.739850 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.849778 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7c8031bb-6ec3-42b3-ace7-80e25071e230-secrets\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.849828 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c8031bb-6ec3-42b3-ace7-80e25071e230-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.849881 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7c8031bb-6ec3-42b3-ace7-80e25071e230-kolla-config\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.849906 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c8031bb-6ec3-42b3-ace7-80e25071e230-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.850029 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7c8031bb-6ec3-42b3-ace7-80e25071e230-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.850074 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7c8031bb-6ec3-42b3-ace7-80e25071e230-config-data-default\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.851622 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.851684 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsvfq\" (UniqueName: \"kubernetes.io/projected/7c8031bb-6ec3-42b3-ace7-80e25071e230-kube-api-access-rsvfq\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.851713 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c8031bb-6ec3-42b3-ace7-80e25071e230-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.954898 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7c8031bb-6ec3-42b3-ace7-80e25071e230-secrets\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.954954 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c8031bb-6ec3-42b3-ace7-80e25071e230-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.955007 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7c8031bb-6ec3-42b3-ace7-80e25071e230-kolla-config\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.955028 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c8031bb-6ec3-42b3-ace7-80e25071e230-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.955070 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7c8031bb-6ec3-42b3-ace7-80e25071e230-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.955096 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7c8031bb-6ec3-42b3-ace7-80e25071e230-config-data-default\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.955118 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.955191 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsvfq\" (UniqueName: \"kubernetes.io/projected/7c8031bb-6ec3-42b3-ace7-80e25071e230-kube-api-access-rsvfq\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.955218 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c8031bb-6ec3-42b3-ace7-80e25071e230-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.956612 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7c8031bb-6ec3-42b3-ace7-80e25071e230-kolla-config\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.956983 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7c8031bb-6ec3-42b3-ace7-80e25071e230-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.957067 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.957727 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7c8031bb-6ec3-42b3-ace7-80e25071e230-config-data-default\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.961409 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c8031bb-6ec3-42b3-ace7-80e25071e230-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.963321 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7c8031bb-6ec3-42b3-ace7-80e25071e230-secrets\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.965291 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c8031bb-6ec3-42b3-ace7-80e25071e230-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.965650 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c8031bb-6ec3-42b3-ace7-80e25071e230-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.977433 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsvfq\" (UniqueName: \"kubernetes.io/projected/7c8031bb-6ec3-42b3-ace7-80e25071e230-kube-api-access-rsvfq\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:47 crc kubenswrapper[4744]: I1003 16:41:47.985163 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"7c8031bb-6ec3-42b3-ace7-80e25071e230\") " pod="openstack/openstack-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.060124 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.074172 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.075601 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.078534 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-qngfh" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.078715 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.078747 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.078952 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.080574 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.158091 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.158156 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.158182 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.158210 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.158237 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvc2s\" (UniqueName: \"kubernetes.io/projected/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-kube-api-access-kvc2s\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.158264 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.158330 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.158358 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.158390 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.259406 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvc2s\" (UniqueName: \"kubernetes.io/projected/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-kube-api-access-kvc2s\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.259451 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.259531 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.259552 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.259578 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.259621 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.259655 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.259675 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.259696 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.260486 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.260630 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.261205 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.261734 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.262865 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.263007 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.263778 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.279150 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvc2s\" (UniqueName: \"kubernetes.io/projected/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-kube-api-access-kvc2s\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.279446 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.292435 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8\") " pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.402375 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.520041 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.521335 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.527629 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.527810 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-8llnh" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.527869 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.531588 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.666254 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7\") " pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.666305 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7smpz\" (UniqueName: \"kubernetes.io/projected/cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7-kube-api-access-7smpz\") pod \"memcached-0\" (UID: \"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7\") " pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.666779 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7\") " pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.666913 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7-config-data\") pod \"memcached-0\" (UID: \"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7\") " pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.667145 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7-kolla-config\") pod \"memcached-0\" (UID: \"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7\") " pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.768443 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7smpz\" (UniqueName: \"kubernetes.io/projected/cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7-kube-api-access-7smpz\") pod \"memcached-0\" (UID: \"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7\") " pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.768562 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7\") " pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.769247 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7\") " pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.769295 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7-config-data\") pod \"memcached-0\" (UID: \"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7\") " pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.769371 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7-kolla-config\") pod \"memcached-0\" (UID: \"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7\") " pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.770207 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7-kolla-config\") pod \"memcached-0\" (UID: \"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7\") " pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.772598 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7\") " pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.773471 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7-config-data\") pod \"memcached-0\" (UID: \"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7\") " pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.774879 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7\") " pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.788773 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7smpz\" (UniqueName: \"kubernetes.io/projected/cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7-kube-api-access-7smpz\") pod \"memcached-0\" (UID: \"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7\") " pod="openstack/memcached-0" Oct 03 16:41:48 crc kubenswrapper[4744]: I1003 16:41:48.843530 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 03 16:41:50 crc kubenswrapper[4744]: I1003 16:41:50.435562 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 16:41:50 crc kubenswrapper[4744]: I1003 16:41:50.437939 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 16:41:50 crc kubenswrapper[4744]: I1003 16:41:50.449279 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-9cjpf" Oct 03 16:41:50 crc kubenswrapper[4744]: I1003 16:41:50.469994 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 16:41:50 crc kubenswrapper[4744]: I1003 16:41:50.501769 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr2f7\" (UniqueName: \"kubernetes.io/projected/e104e42e-cc57-4233-b7a5-2755705d18f7-kube-api-access-cr2f7\") pod \"kube-state-metrics-0\" (UID: \"e104e42e-cc57-4233-b7a5-2755705d18f7\") " pod="openstack/kube-state-metrics-0" Oct 03 16:41:50 crc kubenswrapper[4744]: I1003 16:41:50.603032 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr2f7\" (UniqueName: \"kubernetes.io/projected/e104e42e-cc57-4233-b7a5-2755705d18f7-kube-api-access-cr2f7\") pod \"kube-state-metrics-0\" (UID: \"e104e42e-cc57-4233-b7a5-2755705d18f7\") " pod="openstack/kube-state-metrics-0" Oct 03 16:41:50 crc kubenswrapper[4744]: I1003 16:41:50.631081 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr2f7\" (UniqueName: \"kubernetes.io/projected/e104e42e-cc57-4233-b7a5-2755705d18f7-kube-api-access-cr2f7\") pod \"kube-state-metrics-0\" (UID: \"e104e42e-cc57-4233-b7a5-2755705d18f7\") " pod="openstack/kube-state-metrics-0" Oct 03 16:41:50 crc kubenswrapper[4744]: I1003 16:41:50.770243 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.191282 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4jn8x"] Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.192791 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.195147 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-q4qf7" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.200734 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.200966 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.207045 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4jn8x"] Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.261385 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-var-log-ovn\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.261523 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-var-run\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.261769 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-combined-ca-bundle\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.261846 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-var-run-ovn\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.261951 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47tb5\" (UniqueName: \"kubernetes.io/projected/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-kube-api-access-47tb5\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.262051 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-ovn-controller-tls-certs\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.262093 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-scripts\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.262431 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-99k4w"] Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.263958 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.277823 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-99k4w"] Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.341762 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.346285 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.349966 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.350048 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.351337 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-ndqkc" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.352261 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.353245 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.362270 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.366953 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/77d29fac-2896-4f1a-a193-f4b559f1e9b6-var-log\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.367027 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-combined-ca-bundle\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.367069 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-var-run-ovn\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.367093 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/77d29fac-2896-4f1a-a193-f4b559f1e9b6-var-lib\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.367118 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77d29fac-2896-4f1a-a193-f4b559f1e9b6-scripts\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.367139 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47tb5\" (UniqueName: \"kubernetes.io/projected/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-kube-api-access-47tb5\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.367162 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/77d29fac-2896-4f1a-a193-f4b559f1e9b6-etc-ovs\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.367185 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-ovn-controller-tls-certs\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.367234 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7msb\" (UniqueName: \"kubernetes.io/projected/77d29fac-2896-4f1a-a193-f4b559f1e9b6-kube-api-access-d7msb\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.367261 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/77d29fac-2896-4f1a-a193-f4b559f1e9b6-var-run\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.367284 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-scripts\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.367342 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-var-log-ovn\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.367377 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-var-run\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.367701 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-var-run-ovn\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.367752 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-var-run\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.374792 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-var-log-ovn\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.376813 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-scripts\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.378334 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-combined-ca-bundle\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.378956 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-ovn-controller-tls-certs\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.387529 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47tb5\" (UniqueName: \"kubernetes.io/projected/bc8fb837-d1e5-4953-ae6c-fbee5f867e49-kube-api-access-47tb5\") pod \"ovn-controller-4jn8x\" (UID: \"bc8fb837-d1e5-4953-ae6c-fbee5f867e49\") " pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.470414 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/803d0f2b-9d7d-4d98-909c-18ffdec34aed-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.470479 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.470523 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/803d0f2b-9d7d-4d98-909c-18ffdec34aed-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.470556 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77d29fac-2896-4f1a-a193-f4b559f1e9b6-scripts\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.470594 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/77d29fac-2896-4f1a-a193-f4b559f1e9b6-var-run\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.470673 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mwcq\" (UniqueName: \"kubernetes.io/projected/803d0f2b-9d7d-4d98-909c-18ffdec34aed-kube-api-access-6mwcq\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.470702 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/77d29fac-2896-4f1a-a193-f4b559f1e9b6-var-log\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.470731 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/803d0f2b-9d7d-4d98-909c-18ffdec34aed-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.470764 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/77d29fac-2896-4f1a-a193-f4b559f1e9b6-var-lib\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.470788 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/803d0f2b-9d7d-4d98-909c-18ffdec34aed-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.470813 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/77d29fac-2896-4f1a-a193-f4b559f1e9b6-etc-ovs\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.470822 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/77d29fac-2896-4f1a-a193-f4b559f1e9b6-var-run\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.470839 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/803d0f2b-9d7d-4d98-909c-18ffdec34aed-config\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.470888 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7msb\" (UniqueName: \"kubernetes.io/projected/77d29fac-2896-4f1a-a193-f4b559f1e9b6-kube-api-access-d7msb\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.470917 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/803d0f2b-9d7d-4d98-909c-18ffdec34aed-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.471035 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/77d29fac-2896-4f1a-a193-f4b559f1e9b6-etc-ovs\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.471035 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/77d29fac-2896-4f1a-a193-f4b559f1e9b6-var-log\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.471321 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/77d29fac-2896-4f1a-a193-f4b559f1e9b6-var-lib\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.472725 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77d29fac-2896-4f1a-a193-f4b559f1e9b6-scripts\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.487757 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7msb\" (UniqueName: \"kubernetes.io/projected/77d29fac-2896-4f1a-a193-f4b559f1e9b6-kube-api-access-d7msb\") pod \"ovn-controller-ovs-99k4w\" (UID: \"77d29fac-2896-4f1a-a193-f4b559f1e9b6\") " pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.516004 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4jn8x" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.572794 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/803d0f2b-9d7d-4d98-909c-18ffdec34aed-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.572914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mwcq\" (UniqueName: \"kubernetes.io/projected/803d0f2b-9d7d-4d98-909c-18ffdec34aed-kube-api-access-6mwcq\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.572955 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/803d0f2b-9d7d-4d98-909c-18ffdec34aed-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.572986 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/803d0f2b-9d7d-4d98-909c-18ffdec34aed-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.573007 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/803d0f2b-9d7d-4d98-909c-18ffdec34aed-config\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.573048 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/803d0f2b-9d7d-4d98-909c-18ffdec34aed-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.573084 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/803d0f2b-9d7d-4d98-909c-18ffdec34aed-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.573114 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.574050 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.573943 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/803d0f2b-9d7d-4d98-909c-18ffdec34aed-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.574287 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/803d0f2b-9d7d-4d98-909c-18ffdec34aed-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.574521 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/803d0f2b-9d7d-4d98-909c-18ffdec34aed-config\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.576598 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/803d0f2b-9d7d-4d98-909c-18ffdec34aed-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.581312 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/803d0f2b-9d7d-4d98-909c-18ffdec34aed-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.588090 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/803d0f2b-9d7d-4d98-909c-18ffdec34aed-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.590845 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.597866 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mwcq\" (UniqueName: \"kubernetes.io/projected/803d0f2b-9d7d-4d98-909c-18ffdec34aed-kube-api-access-6mwcq\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.610786 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"803d0f2b-9d7d-4d98-909c-18ffdec34aed\") " pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:54 crc kubenswrapper[4744]: I1003 16:41:54.674244 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 03 16:41:55 crc kubenswrapper[4744]: E1003 16:41:55.947903 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 03 16:41:55 crc kubenswrapper[4744]: E1003 16:41:55.948303 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fvhgq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-hvq55_openstack(9d8609ec-6619-4d44-a1f2-3b91146766ba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 16:41:55 crc kubenswrapper[4744]: E1003 16:41:55.949455 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-hvq55" podUID="9d8609ec-6619-4d44-a1f2-3b91146766ba" Oct 03 16:41:56 crc kubenswrapper[4744]: E1003 16:41:56.224196 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 03 16:41:56 crc kubenswrapper[4744]: E1003 16:41:56.224508 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gsgnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-6ksgr_openstack(7e8436a4-6ba0-4112-84ea-4c40491cea31): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 16:41:56 crc kubenswrapper[4744]: E1003 16:41:56.226281 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" podUID="7e8436a4-6ba0-4112-84ea-4c40491cea31" Oct 03 16:41:56 crc kubenswrapper[4744]: I1003 16:41:56.342528 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-hvq55" Oct 03 16:41:56 crc kubenswrapper[4744]: I1003 16:41:56.415116 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8609ec-6619-4d44-a1f2-3b91146766ba-config\") pod \"9d8609ec-6619-4d44-a1f2-3b91146766ba\" (UID: \"9d8609ec-6619-4d44-a1f2-3b91146766ba\") " Oct 03 16:41:56 crc kubenswrapper[4744]: I1003 16:41:56.415645 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvhgq\" (UniqueName: \"kubernetes.io/projected/9d8609ec-6619-4d44-a1f2-3b91146766ba-kube-api-access-fvhgq\") pod \"9d8609ec-6619-4d44-a1f2-3b91146766ba\" (UID: \"9d8609ec-6619-4d44-a1f2-3b91146766ba\") " Oct 03 16:41:56 crc kubenswrapper[4744]: I1003 16:41:56.416433 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d8609ec-6619-4d44-a1f2-3b91146766ba-config" (OuterVolumeSpecName: "config") pod "9d8609ec-6619-4d44-a1f2-3b91146766ba" (UID: "9d8609ec-6619-4d44-a1f2-3b91146766ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:41:56 crc kubenswrapper[4744]: I1003 16:41:56.423466 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d8609ec-6619-4d44-a1f2-3b91146766ba-kube-api-access-fvhgq" (OuterVolumeSpecName: "kube-api-access-fvhgq") pod "9d8609ec-6619-4d44-a1f2-3b91146766ba" (UID: "9d8609ec-6619-4d44-a1f2-3b91146766ba"). InnerVolumeSpecName "kube-api-access-fvhgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:41:56 crc kubenswrapper[4744]: I1003 16:41:56.518379 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvhgq\" (UniqueName: \"kubernetes.io/projected/9d8609ec-6619-4d44-a1f2-3b91146766ba-kube-api-access-fvhgq\") on node \"crc\" DevicePath \"\"" Oct 03 16:41:56 crc kubenswrapper[4744]: I1003 16:41:56.518686 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8609ec-6619-4d44-a1f2-3b91146766ba-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:41:56 crc kubenswrapper[4744]: I1003 16:41:56.791406 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 16:41:56 crc kubenswrapper[4744]: I1003 16:41:56.796920 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4jn8x"] Oct 03 16:41:56 crc kubenswrapper[4744]: I1003 16:41:56.807269 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 16:41:56 crc kubenswrapper[4744]: W1003 16:41:56.846677 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda05294a3_e8e1_40cb_8e34_1c4fd4f0c533.slice/crio-fc147916d7be3ea0ecffcea13c3393b9a48fef121fe2ca2d7a881352ea627628 WatchSource:0}: Error finding container fc147916d7be3ea0ecffcea13c3393b9a48fef121fe2ca2d7a881352ea627628: Status 404 returned error can't find the container with id fc147916d7be3ea0ecffcea13c3393b9a48fef121fe2ca2d7a881352ea627628 Oct 03 16:41:56 crc kubenswrapper[4744]: I1003 16:41:56.947719 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 03 16:41:56 crc kubenswrapper[4744]: W1003 16:41:56.956697 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod803d0f2b_9d7d_4d98_909c_18ffdec34aed.slice/crio-ca4b4e7bac51648f5cdc46797b6511c7c012933960d5275e29fa64ccd25be072 WatchSource:0}: Error finding container ca4b4e7bac51648f5cdc46797b6511c7c012933960d5275e29fa64ccd25be072: Status 404 returned error can't find the container with id ca4b4e7bac51648f5cdc46797b6511c7c012933960d5275e29fa64ccd25be072 Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.001396 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.019534 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 03 16:41:57 crc kubenswrapper[4744]: W1003 16:41:57.025289 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode104e42e_cc57_4233_b7a5_2755705d18f7.slice/crio-983936e1ca08042b3c479d3bd4e40443aa75778f55c0eba77cb4e5ccdaca0de0 WatchSource:0}: Error finding container 983936e1ca08042b3c479d3bd4e40443aa75778f55c0eba77cb4e5ccdaca0de0: Status 404 returned error can't find the container with id 983936e1ca08042b3c479d3bd4e40443aa75778f55c0eba77cb4e5ccdaca0de0 Oct 03 16:41:57 crc kubenswrapper[4744]: W1003 16:41:57.026436 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71ec1585_ca83_4fd6_9bb2_f6e92b29cbe8.slice/crio-ecb402dd482a2780fd9d8137dba2a7eaa4b253f0ef99bf60d7390b3662b4ac5c WatchSource:0}: Error finding container ecb402dd482a2780fd9d8137dba2a7eaa4b253f0ef99bf60d7390b3662b4ac5c: Status 404 returned error can't find the container with id ecb402dd482a2780fd9d8137dba2a7eaa4b253f0ef99bf60d7390b3662b4ac5c Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.027801 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4jn8x" event={"ID":"bc8fb837-d1e5-4953-ae6c-fbee5f867e49","Type":"ContainerStarted","Data":"5f0b7e52172b8ddd896bc7731470acfa924d6830f6f37f8dd6ddb3d2f0f7c1df"} Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.029658 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-hvq55" event={"ID":"9d8609ec-6619-4d44-a1f2-3b91146766ba","Type":"ContainerDied","Data":"aa517e0ffbd19e44ee8b04aed7e03721b8f5eb9693aeeefcec699703908fb0e4"} Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.029728 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-hvq55" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.033984 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533","Type":"ContainerStarted","Data":"fc147916d7be3ea0ecffcea13c3393b9a48fef121fe2ca2d7a881352ea627628"} Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.039800 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"803d0f2b-9d7d-4d98-909c-18ffdec34aed","Type":"ContainerStarted","Data":"ca4b4e7bac51648f5cdc46797b6511c7c012933960d5275e29fa64ccd25be072"} Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.043835 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.051470 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-64zgj"] Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.060615 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bwpkl"] Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.077208 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d21c39f-eece-4cf0-847b-c8ebc87c574d","Type":"ContainerStarted","Data":"4dfe06a0aafe99f6a5ab07830dda957bac86755dfd75c69bca2377b1473d75a6"} Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.094027 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-hvq55"] Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.104769 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-hvq55"] Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.145578 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-99k4w"] Oct 03 16:41:57 crc kubenswrapper[4744]: W1003 16:41:57.154955 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77d29fac_2896_4f1a_a193_f4b559f1e9b6.slice/crio-a876da2d5ccf38930f6b2cf31b01d31a510008c4d77c16f465711f061fccaca1 WatchSource:0}: Error finding container a876da2d5ccf38930f6b2cf31b01d31a510008c4d77c16f465711f061fccaca1: Status 404 returned error can't find the container with id a876da2d5ccf38930f6b2cf31b01d31a510008c4d77c16f465711f061fccaca1 Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.208222 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 03 16:41:57 crc kubenswrapper[4744]: W1003 16:41:57.216701 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc1aa9df_ee4d_4cd6_b35a_914b9656c0b7.slice/crio-42a1ccbe8e5f0cc34ad36b508589012d4ad6aa39d99d579f6474767e3d559eb6 WatchSource:0}: Error finding container 42a1ccbe8e5f0cc34ad36b508589012d4ad6aa39d99d579f6474767e3d559eb6: Status 404 returned error can't find the container with id 42a1ccbe8e5f0cc34ad36b508589012d4ad6aa39d99d579f6474767e3d559eb6 Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.322677 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.432326 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e8436a4-6ba0-4112-84ea-4c40491cea31-config\") pod \"7e8436a4-6ba0-4112-84ea-4c40491cea31\" (UID: \"7e8436a4-6ba0-4112-84ea-4c40491cea31\") " Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.432748 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e8436a4-6ba0-4112-84ea-4c40491cea31-dns-svc\") pod \"7e8436a4-6ba0-4112-84ea-4c40491cea31\" (UID: \"7e8436a4-6ba0-4112-84ea-4c40491cea31\") " Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.432849 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsgnp\" (UniqueName: \"kubernetes.io/projected/7e8436a4-6ba0-4112-84ea-4c40491cea31-kube-api-access-gsgnp\") pod \"7e8436a4-6ba0-4112-84ea-4c40491cea31\" (UID: \"7e8436a4-6ba0-4112-84ea-4c40491cea31\") " Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.432901 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e8436a4-6ba0-4112-84ea-4c40491cea31-config" (OuterVolumeSpecName: "config") pod "7e8436a4-6ba0-4112-84ea-4c40491cea31" (UID: "7e8436a4-6ba0-4112-84ea-4c40491cea31"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.433296 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e8436a4-6ba0-4112-84ea-4c40491cea31-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.433305 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e8436a4-6ba0-4112-84ea-4c40491cea31-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7e8436a4-6ba0-4112-84ea-4c40491cea31" (UID: "7e8436a4-6ba0-4112-84ea-4c40491cea31"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.439288 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e8436a4-6ba0-4112-84ea-4c40491cea31-kube-api-access-gsgnp" (OuterVolumeSpecName: "kube-api-access-gsgnp") pod "7e8436a4-6ba0-4112-84ea-4c40491cea31" (UID: "7e8436a4-6ba0-4112-84ea-4c40491cea31"). InnerVolumeSpecName "kube-api-access-gsgnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.535114 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e8436a4-6ba0-4112-84ea-4c40491cea31-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.535150 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsgnp\" (UniqueName: \"kubernetes.io/projected/7e8436a4-6ba0-4112-84ea-4c40491cea31-kube-api-access-gsgnp\") on node \"crc\" DevicePath \"\"" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.721370 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.723233 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.724694 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-h6bzj" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.727433 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.727619 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.728135 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.750143 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.848209 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c9e2064-beda-46d7-9221-06b2bbb375eb-config\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.848280 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c9e2064-beda-46d7-9221-06b2bbb375eb-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.848353 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c9e2064-beda-46d7-9221-06b2bbb375eb-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.848391 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c9e2064-beda-46d7-9221-06b2bbb375eb-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.848423 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.848455 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c9e2064-beda-46d7-9221-06b2bbb375eb-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.848475 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c9e2064-beda-46d7-9221-06b2bbb375eb-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.848535 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79kf6\" (UniqueName: \"kubernetes.io/projected/7c9e2064-beda-46d7-9221-06b2bbb375eb-kube-api-access-79kf6\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.949741 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79kf6\" (UniqueName: \"kubernetes.io/projected/7c9e2064-beda-46d7-9221-06b2bbb375eb-kube-api-access-79kf6\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.949814 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c9e2064-beda-46d7-9221-06b2bbb375eb-config\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.949834 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c9e2064-beda-46d7-9221-06b2bbb375eb-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.949876 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c9e2064-beda-46d7-9221-06b2bbb375eb-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.949912 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c9e2064-beda-46d7-9221-06b2bbb375eb-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.949942 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.949982 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c9e2064-beda-46d7-9221-06b2bbb375eb-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.950005 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c9e2064-beda-46d7-9221-06b2bbb375eb-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.952594 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.952739 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7c9e2064-beda-46d7-9221-06b2bbb375eb-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.953442 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c9e2064-beda-46d7-9221-06b2bbb375eb-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.953455 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c9e2064-beda-46d7-9221-06b2bbb375eb-config\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.955291 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c9e2064-beda-46d7-9221-06b2bbb375eb-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.955795 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c9e2064-beda-46d7-9221-06b2bbb375eb-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.957764 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7c9e2064-beda-46d7-9221-06b2bbb375eb-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.968940 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79kf6\" (UniqueName: \"kubernetes.io/projected/7c9e2064-beda-46d7-9221-06b2bbb375eb-kube-api-access-79kf6\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:57 crc kubenswrapper[4744]: I1003 16:41:57.985766 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7c9e2064-beda-46d7-9221-06b2bbb375eb\") " pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.094681 4744 generic.go:334] "Generic (PLEG): container finished" podID="ccc57209-f9dc-4308-b11e-d3e10c70fe0a" containerID="bc3cc215bec8054bff3ce779162950ce5320408680bf2ba3e8a96355d9b6785e" exitCode=0 Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.094764 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-64zgj" event={"ID":"ccc57209-f9dc-4308-b11e-d3e10c70fe0a","Type":"ContainerDied","Data":"bc3cc215bec8054bff3ce779162950ce5320408680bf2ba3e8a96355d9b6785e"} Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.094823 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-64zgj" event={"ID":"ccc57209-f9dc-4308-b11e-d3e10c70fe0a","Type":"ContainerStarted","Data":"4613b61d2c7139f9bf97adae63885d82b2903577a043ecb231c44276b0c58e97"} Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.098295 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e104e42e-cc57-4233-b7a5-2755705d18f7","Type":"ContainerStarted","Data":"983936e1ca08042b3c479d3bd4e40443aa75778f55c0eba77cb4e5ccdaca0de0"} Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.100143 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8","Type":"ContainerStarted","Data":"ecb402dd482a2780fd9d8137dba2a7eaa4b253f0ef99bf60d7390b3662b4ac5c"} Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.101920 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7","Type":"ContainerStarted","Data":"42a1ccbe8e5f0cc34ad36b508589012d4ad6aa39d99d579f6474767e3d559eb6"} Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.107120 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-99k4w" event={"ID":"77d29fac-2896-4f1a-a193-f4b559f1e9b6","Type":"ContainerStarted","Data":"a876da2d5ccf38930f6b2cf31b01d31a510008c4d77c16f465711f061fccaca1"} Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.111535 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7c8031bb-6ec3-42b3-ace7-80e25071e230","Type":"ContainerStarted","Data":"1fd2d1dd3bca9c8b69e53a468f6e5a70d633d185b5016748c3061f99a4cd1518"} Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.116276 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" event={"ID":"7e8436a4-6ba0-4112-84ea-4c40491cea31","Type":"ContainerDied","Data":"fcffff7dd791184eff9d72e347ed1ec14a2c985e5c1950d45a519dbabd66630f"} Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.116327 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6ksgr" Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.117978 4744 generic.go:334] "Generic (PLEG): container finished" podID="19d490cc-8f49-4957-9812-94e96b24e100" containerID="4f1d7cbce0d835fa92818618aa79a0fd1f46830c20db70bea380e4c2cc675c65" exitCode=0 Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.118001 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" event={"ID":"19d490cc-8f49-4957-9812-94e96b24e100","Type":"ContainerDied","Data":"4f1d7cbce0d835fa92818618aa79a0fd1f46830c20db70bea380e4c2cc675c65"} Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.118015 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" event={"ID":"19d490cc-8f49-4957-9812-94e96b24e100","Type":"ContainerStarted","Data":"3d05c868402777cc598a7982adfa17791e6feaa03e807a7cb7049d72b6af3414"} Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.124124 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.191519 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6ksgr"] Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.198774 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6ksgr"] Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.902658 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e8436a4-6ba0-4112-84ea-4c40491cea31" path="/var/lib/kubelet/pods/7e8436a4-6ba0-4112-84ea-4c40491cea31/volumes" Oct 03 16:41:58 crc kubenswrapper[4744]: I1003 16:41:58.903141 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d8609ec-6619-4d44-a1f2-3b91146766ba" path="/var/lib/kubelet/pods/9d8609ec-6619-4d44-a1f2-3b91146766ba/volumes" Oct 03 16:42:07 crc kubenswrapper[4744]: I1003 16:42:07.538225 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 03 16:42:07 crc kubenswrapper[4744]: W1003 16:42:07.675985 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c9e2064_beda_46d7_9221_06b2bbb375eb.slice/crio-dcfdd48557e6c945a385215919827cf9f7768abebd2fec03e555084cd51f88e5 WatchSource:0}: Error finding container dcfdd48557e6c945a385215919827cf9f7768abebd2fec03e555084cd51f88e5: Status 404 returned error can't find the container with id dcfdd48557e6c945a385215919827cf9f7768abebd2fec03e555084cd51f88e5 Oct 03 16:42:08 crc kubenswrapper[4744]: I1003 16:42:08.206424 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" event={"ID":"19d490cc-8f49-4957-9812-94e96b24e100","Type":"ContainerStarted","Data":"4adc4530437a2232f9e63b1279dafeabf9af51275becdfd5424943bd1643b973"} Oct 03 16:42:08 crc kubenswrapper[4744]: I1003 16:42:08.207069 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" Oct 03 16:42:08 crc kubenswrapper[4744]: I1003 16:42:08.209324 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7c9e2064-beda-46d7-9221-06b2bbb375eb","Type":"ContainerStarted","Data":"dcfdd48557e6c945a385215919827cf9f7768abebd2fec03e555084cd51f88e5"} Oct 03 16:42:08 crc kubenswrapper[4744]: I1003 16:42:08.212888 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-64zgj" event={"ID":"ccc57209-f9dc-4308-b11e-d3e10c70fe0a","Type":"ContainerStarted","Data":"2e5acad8a02a926b49ce792a513128c48289fccaeb3b94d4950d6370278f4373"} Oct 03 16:42:08 crc kubenswrapper[4744]: I1003 16:42:08.212982 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-64zgj" Oct 03 16:42:08 crc kubenswrapper[4744]: I1003 16:42:08.215696 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7","Type":"ContainerStarted","Data":"e24b95a3d7ee62cf35176898bb2b45c802db685e45db7b67c7e2a294ca65fba1"} Oct 03 16:42:08 crc kubenswrapper[4744]: I1003 16:42:08.216406 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 03 16:42:08 crc kubenswrapper[4744]: I1003 16:42:08.228275 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" podStartSLOduration=23.769965002 podStartE2EDuration="24.228250843s" podCreationTimestamp="2025-10-03 16:41:44 +0000 UTC" firstStartedPulling="2025-10-03 16:41:57.079042757 +0000 UTC m=+1043.358918653" lastFinishedPulling="2025-10-03 16:41:57.537328598 +0000 UTC m=+1043.817204494" observedRunningTime="2025-10-03 16:42:08.226255341 +0000 UTC m=+1054.506131237" watchObservedRunningTime="2025-10-03 16:42:08.228250843 +0000 UTC m=+1054.508126739" Oct 03 16:42:08 crc kubenswrapper[4744]: I1003 16:42:08.245004 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-64zgj" podStartSLOduration=23.71799826 podStartE2EDuration="24.244985235s" podCreationTimestamp="2025-10-03 16:41:44 +0000 UTC" firstStartedPulling="2025-10-03 16:41:57.077542728 +0000 UTC m=+1043.357418624" lastFinishedPulling="2025-10-03 16:41:57.604529703 +0000 UTC m=+1043.884405599" observedRunningTime="2025-10-03 16:42:08.241506185 +0000 UTC m=+1054.521382091" watchObservedRunningTime="2025-10-03 16:42:08.244985235 +0000 UTC m=+1054.524861131" Oct 03 16:42:08 crc kubenswrapper[4744]: I1003 16:42:08.262462 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=10.633832604 podStartE2EDuration="20.262439085s" podCreationTimestamp="2025-10-03 16:41:48 +0000 UTC" firstStartedPulling="2025-10-03 16:41:57.248530412 +0000 UTC m=+1043.528406308" lastFinishedPulling="2025-10-03 16:42:06.877136893 +0000 UTC m=+1053.157012789" observedRunningTime="2025-10-03 16:42:08.257239151 +0000 UTC m=+1054.537115067" watchObservedRunningTime="2025-10-03 16:42:08.262439085 +0000 UTC m=+1054.542314981" Oct 03 16:42:09 crc kubenswrapper[4744]: I1003 16:42:09.224905 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7c8031bb-6ec3-42b3-ace7-80e25071e230","Type":"ContainerStarted","Data":"10551ff23f2769eeac97f46e609b56e94369ffa84f5fbc5804495ae9d1ee8047"} Oct 03 16:42:09 crc kubenswrapper[4744]: I1003 16:42:09.226823 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"803d0f2b-9d7d-4d98-909c-18ffdec34aed","Type":"ContainerStarted","Data":"e58a3cfb2a0aa8d86502297768f19448408bee3223bdee99f3d226343f196af9"} Oct 03 16:42:09 crc kubenswrapper[4744]: I1003 16:42:09.228754 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e104e42e-cc57-4233-b7a5-2755705d18f7","Type":"ContainerStarted","Data":"3090f3f9894d8ef9f8d9519dbde832d68952a7ea5dc64f2385cdfa0d661cd64e"} Oct 03 16:42:09 crc kubenswrapper[4744]: I1003 16:42:09.228907 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 03 16:42:09 crc kubenswrapper[4744]: I1003 16:42:09.230804 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4jn8x" event={"ID":"bc8fb837-d1e5-4953-ae6c-fbee5f867e49","Type":"ContainerStarted","Data":"443d99ed105721b951afcabe3f914c622cfea5f296f6bf84f48fe72e7ee142fd"} Oct 03 16:42:09 crc kubenswrapper[4744]: I1003 16:42:09.230927 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-4jn8x" Oct 03 16:42:09 crc kubenswrapper[4744]: I1003 16:42:09.232465 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8","Type":"ContainerStarted","Data":"df47ee35793173be25cfdf3e663c07efe415d1001af848235abedc3366df7261"} Oct 03 16:42:09 crc kubenswrapper[4744]: I1003 16:42:09.234025 4744 generic.go:334] "Generic (PLEG): container finished" podID="77d29fac-2896-4f1a-a193-f4b559f1e9b6" containerID="7bb487d771a101eab448fef6e5a0ec23c11c822122f334f4d44dd5d7337ad641" exitCode=0 Oct 03 16:42:09 crc kubenswrapper[4744]: I1003 16:42:09.234083 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-99k4w" event={"ID":"77d29fac-2896-4f1a-a193-f4b559f1e9b6","Type":"ContainerDied","Data":"7bb487d771a101eab448fef6e5a0ec23c11c822122f334f4d44dd5d7337ad641"} Oct 03 16:42:09 crc kubenswrapper[4744]: I1003 16:42:09.292812 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=8.512492372 podStartE2EDuration="19.292794955s" podCreationTimestamp="2025-10-03 16:41:50 +0000 UTC" firstStartedPulling="2025-10-03 16:41:57.077451375 +0000 UTC m=+1043.357327271" lastFinishedPulling="2025-10-03 16:42:07.857753948 +0000 UTC m=+1054.137629854" observedRunningTime="2025-10-03 16:42:09.28560434 +0000 UTC m=+1055.565480226" watchObservedRunningTime="2025-10-03 16:42:09.292794955 +0000 UTC m=+1055.572670851" Oct 03 16:42:09 crc kubenswrapper[4744]: I1003 16:42:09.317916 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-4jn8x" podStartSLOduration=5.053074038 podStartE2EDuration="15.317894803s" podCreationTimestamp="2025-10-03 16:41:54 +0000 UTC" firstStartedPulling="2025-10-03 16:41:56.809877829 +0000 UTC m=+1043.089753725" lastFinishedPulling="2025-10-03 16:42:07.074698594 +0000 UTC m=+1053.354574490" observedRunningTime="2025-10-03 16:42:09.312886594 +0000 UTC m=+1055.592762690" watchObservedRunningTime="2025-10-03 16:42:09.317894803 +0000 UTC m=+1055.597770699" Oct 03 16:42:10 crc kubenswrapper[4744]: I1003 16:42:10.245551 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-99k4w" event={"ID":"77d29fac-2896-4f1a-a193-f4b559f1e9b6","Type":"ContainerStarted","Data":"dd554eb5955fe5e73ee838c8de1af9669fcdca28fdff300dc2841f526be94e8e"} Oct 03 16:42:10 crc kubenswrapper[4744]: I1003 16:42:10.246124 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:42:10 crc kubenswrapper[4744]: I1003 16:42:10.246139 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-99k4w" event={"ID":"77d29fac-2896-4f1a-a193-f4b559f1e9b6","Type":"ContainerStarted","Data":"74c16268b3556c5d025ce37d02fc28f5a2fd67662fe926c46353d6c4f173a34b"} Oct 03 16:42:10 crc kubenswrapper[4744]: I1003 16:42:10.246153 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:42:10 crc kubenswrapper[4744]: I1003 16:42:10.249253 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533","Type":"ContainerStarted","Data":"468e9fb03dafde6384d03462697979e76f31467f870859d2320a59286b385cf5"} Oct 03 16:42:10 crc kubenswrapper[4744]: I1003 16:42:10.251454 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7c9e2064-beda-46d7-9221-06b2bbb375eb","Type":"ContainerStarted","Data":"3988ffe5d6f0ad33b44413d4543f5498ad505cc485542e9dc1feec3e83fdcb46"} Oct 03 16:42:10 crc kubenswrapper[4744]: I1003 16:42:10.255433 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d21c39f-eece-4cf0-847b-c8ebc87c574d","Type":"ContainerStarted","Data":"50822061c1b75d267e0d3c6f99f1d88af5ebf5ff228151acf820e4a5ab291a81"} Oct 03 16:42:10 crc kubenswrapper[4744]: I1003 16:42:10.273034 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-99k4w" podStartSLOduration=6.233115412 podStartE2EDuration="16.273009581s" podCreationTimestamp="2025-10-03 16:41:54 +0000 UTC" firstStartedPulling="2025-10-03 16:41:57.157563734 +0000 UTC m=+1043.437439630" lastFinishedPulling="2025-10-03 16:42:07.197457903 +0000 UTC m=+1053.477333799" observedRunningTime="2025-10-03 16:42:10.265484986 +0000 UTC m=+1056.545360902" watchObservedRunningTime="2025-10-03 16:42:10.273009581 +0000 UTC m=+1056.552885477" Oct 03 16:42:12 crc kubenswrapper[4744]: I1003 16:42:12.325702 4744 generic.go:334] "Generic (PLEG): container finished" podID="71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8" containerID="df47ee35793173be25cfdf3e663c07efe415d1001af848235abedc3366df7261" exitCode=0 Oct 03 16:42:12 crc kubenswrapper[4744]: I1003 16:42:12.325790 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8","Type":"ContainerDied","Data":"df47ee35793173be25cfdf3e663c07efe415d1001af848235abedc3366df7261"} Oct 03 16:42:12 crc kubenswrapper[4744]: I1003 16:42:12.328620 4744 generic.go:334] "Generic (PLEG): container finished" podID="7c8031bb-6ec3-42b3-ace7-80e25071e230" containerID="10551ff23f2769eeac97f46e609b56e94369ffa84f5fbc5804495ae9d1ee8047" exitCode=0 Oct 03 16:42:12 crc kubenswrapper[4744]: I1003 16:42:12.328703 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7c8031bb-6ec3-42b3-ace7-80e25071e230","Type":"ContainerDied","Data":"10551ff23f2769eeac97f46e609b56e94369ffa84f5fbc5804495ae9d1ee8047"} Oct 03 16:42:12 crc kubenswrapper[4744]: I1003 16:42:12.332652 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7c9e2064-beda-46d7-9221-06b2bbb375eb","Type":"ContainerStarted","Data":"21b4264f7080452de15a1e67bb13033b819f07c9307ccd26fd9ff06ce69356f4"} Oct 03 16:42:12 crc kubenswrapper[4744]: I1003 16:42:12.338246 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"803d0f2b-9d7d-4d98-909c-18ffdec34aed","Type":"ContainerStarted","Data":"a920e7287adcf4749ed2b9cee6d32f1675d807cec28750165cbdc57e945cbb67"} Oct 03 16:42:12 crc kubenswrapper[4744]: I1003 16:42:12.413055 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.836293997 podStartE2EDuration="19.413025697s" podCreationTimestamp="2025-10-03 16:41:53 +0000 UTC" firstStartedPulling="2025-10-03 16:41:56.96256359 +0000 UTC m=+1043.242439486" lastFinishedPulling="2025-10-03 16:42:11.53929529 +0000 UTC m=+1057.819171186" observedRunningTime="2025-10-03 16:42:12.404436925 +0000 UTC m=+1058.684312821" watchObservedRunningTime="2025-10-03 16:42:12.413025697 +0000 UTC m=+1058.692901593" Oct 03 16:42:12 crc kubenswrapper[4744]: I1003 16:42:12.429487 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=12.584833079 podStartE2EDuration="16.429460941s" podCreationTimestamp="2025-10-03 16:41:56 +0000 UTC" firstStartedPulling="2025-10-03 16:42:07.681603771 +0000 UTC m=+1053.961479667" lastFinishedPulling="2025-10-03 16:42:11.526231633 +0000 UTC m=+1057.806107529" observedRunningTime="2025-10-03 16:42:12.422892601 +0000 UTC m=+1058.702768507" watchObservedRunningTime="2025-10-03 16:42:12.429460941 +0000 UTC m=+1058.709336837" Oct 03 16:42:12 crc kubenswrapper[4744]: I1003 16:42:12.677414 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 03 16:42:12 crc kubenswrapper[4744]: I1003 16:42:12.714648 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.124450 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.124564 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.169377 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.347766 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8","Type":"ContainerStarted","Data":"f012b20d08b3d08c40e583de387886fce098a455a35fbe405145a3a8c01f2a7d"} Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.350166 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7c8031bb-6ec3-42b3-ace7-80e25071e230","Type":"ContainerStarted","Data":"86539f31b29495fa9993f67b8040d2af9e142fd20143130dd8e22676cc74b04b"} Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.351028 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.370300 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=16.522024488 podStartE2EDuration="26.37028004s" podCreationTimestamp="2025-10-03 16:41:47 +0000 UTC" firstStartedPulling="2025-10-03 16:41:57.029394945 +0000 UTC m=+1043.309270851" lastFinishedPulling="2025-10-03 16:42:06.877650507 +0000 UTC m=+1053.157526403" observedRunningTime="2025-10-03 16:42:13.370093005 +0000 UTC m=+1059.649968961" watchObservedRunningTime="2025-10-03 16:42:13.37028004 +0000 UTC m=+1059.650155936" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.403954 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=16.750100251 podStartE2EDuration="27.403442646s" podCreationTimestamp="2025-10-03 16:41:46 +0000 UTC" firstStartedPulling="2025-10-03 16:41:57.077666091 +0000 UTC m=+1043.357541987" lastFinishedPulling="2025-10-03 16:42:07.731008486 +0000 UTC m=+1054.010884382" observedRunningTime="2025-10-03 16:42:13.396347233 +0000 UTC m=+1059.676223209" watchObservedRunningTime="2025-10-03 16:42:13.403442646 +0000 UTC m=+1059.683318562" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.433625 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.704115 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bwpkl"] Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.704775 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" podUID="19d490cc-8f49-4957-9812-94e96b24e100" containerName="dnsmasq-dns" containerID="cri-o://4adc4530437a2232f9e63b1279dafeabf9af51275becdfd5424943bd1643b973" gracePeriod=10 Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.708065 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.761683 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-mgnmg"] Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.763047 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.768877 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.778029 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-tr4vq"] Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.784786 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.787932 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.814106 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-mgnmg"] Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.828871 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-tr4vq"] Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.846252 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-config\") pod \"dnsmasq-dns-7fd796d7df-tr4vq\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.846360 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/02bac763-b977-4d19-a6fd-300b0b2cefc3-ovn-rundir\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.846439 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/02bac763-b977-4d19-a6fd-300b0b2cefc3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.846636 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02bac763-b977-4d19-a6fd-300b0b2cefc3-combined-ca-bundle\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.846716 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q95m\" (UniqueName: \"kubernetes.io/projected/02bac763-b977-4d19-a6fd-300b0b2cefc3-kube-api-access-7q95m\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.846790 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42wfv\" (UniqueName: \"kubernetes.io/projected/cf8ea9a9-38eb-4588-82ef-831abd8841cc-kube-api-access-42wfv\") pod \"dnsmasq-dns-7fd796d7df-tr4vq\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.846859 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02bac763-b977-4d19-a6fd-300b0b2cefc3-config\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.846937 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-tr4vq\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.846980 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-tr4vq\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.847209 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/02bac763-b977-4d19-a6fd-300b0b2cefc3-ovs-rundir\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.856455 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.950048 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02bac763-b977-4d19-a6fd-300b0b2cefc3-combined-ca-bundle\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.950101 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q95m\" (UniqueName: \"kubernetes.io/projected/02bac763-b977-4d19-a6fd-300b0b2cefc3-kube-api-access-7q95m\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.950119 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42wfv\" (UniqueName: \"kubernetes.io/projected/cf8ea9a9-38eb-4588-82ef-831abd8841cc-kube-api-access-42wfv\") pod \"dnsmasq-dns-7fd796d7df-tr4vq\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.950143 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02bac763-b977-4d19-a6fd-300b0b2cefc3-config\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.950170 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-tr4vq\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.950190 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-tr4vq\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.950266 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/02bac763-b977-4d19-a6fd-300b0b2cefc3-ovs-rundir\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.950303 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-config\") pod \"dnsmasq-dns-7fd796d7df-tr4vq\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.950322 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/02bac763-b977-4d19-a6fd-300b0b2cefc3-ovn-rundir\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.950344 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/02bac763-b977-4d19-a6fd-300b0b2cefc3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.953244 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/02bac763-b977-4d19-a6fd-300b0b2cefc3-ovs-rundir\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.953650 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/02bac763-b977-4d19-a6fd-300b0b2cefc3-ovn-rundir\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.954117 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-tr4vq\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.954231 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02bac763-b977-4d19-a6fd-300b0b2cefc3-config\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.954703 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-config\") pod \"dnsmasq-dns-7fd796d7df-tr4vq\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.955422 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-tr4vq\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.971342 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02bac763-b977-4d19-a6fd-300b0b2cefc3-combined-ca-bundle\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.979084 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q95m\" (UniqueName: \"kubernetes.io/projected/02bac763-b977-4d19-a6fd-300b0b2cefc3-kube-api-access-7q95m\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:13 crc kubenswrapper[4744]: I1003 16:42:13.988797 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/02bac763-b977-4d19-a6fd-300b0b2cefc3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-mgnmg\" (UID: \"02bac763-b977-4d19-a6fd-300b0b2cefc3\") " pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.033371 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42wfv\" (UniqueName: \"kubernetes.io/projected/cf8ea9a9-38eb-4588-82ef-831abd8841cc-kube-api-access-42wfv\") pod \"dnsmasq-dns-7fd796d7df-tr4vq\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.082132 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-64zgj"] Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.082378 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-64zgj" podUID="ccc57209-f9dc-4308-b11e-d3e10c70fe0a" containerName="dnsmasq-dns" containerID="cri-o://2e5acad8a02a926b49ce792a513128c48289fccaeb3b94d4950d6370278f4373" gracePeriod=10 Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.089837 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-64zgj" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.157651 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9qqd6"] Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.159630 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-mgnmg" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.167866 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.169861 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.187979 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.201257 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9qqd6"] Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.265207 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-9qqd6\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.265489 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m2f7\" (UniqueName: \"kubernetes.io/projected/39bde481-e439-403f-8b03-8775c82ced64-kube-api-access-8m2f7\") pod \"dnsmasq-dns-86db49b7ff-9qqd6\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.265547 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-9qqd6\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.265564 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-9qqd6\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.265623 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-config\") pod \"dnsmasq-dns-86db49b7ff-9qqd6\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.369807 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-9qqd6\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.369842 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m2f7\" (UniqueName: \"kubernetes.io/projected/39bde481-e439-403f-8b03-8775c82ced64-kube-api-access-8m2f7\") pod \"dnsmasq-dns-86db49b7ff-9qqd6\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.369884 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-9qqd6\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.369901 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-9qqd6\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.369968 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-config\") pod \"dnsmasq-dns-86db49b7ff-9qqd6\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.370914 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-config\") pod \"dnsmasq-dns-86db49b7ff-9qqd6\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.371423 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-9qqd6\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.372353 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-9qqd6\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.372956 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-9qqd6\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.400077 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m2f7\" (UniqueName: \"kubernetes.io/projected/39bde481-e439-403f-8b03-8775c82ced64-kube-api-access-8m2f7\") pod \"dnsmasq-dns-86db49b7ff-9qqd6\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.417263 4744 generic.go:334] "Generic (PLEG): container finished" podID="ccc57209-f9dc-4308-b11e-d3e10c70fe0a" containerID="2e5acad8a02a926b49ce792a513128c48289fccaeb3b94d4950d6370278f4373" exitCode=0 Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.417535 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-64zgj" event={"ID":"ccc57209-f9dc-4308-b11e-d3e10c70fe0a","Type":"ContainerDied","Data":"2e5acad8a02a926b49ce792a513128c48289fccaeb3b94d4950d6370278f4373"} Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.426836 4744 generic.go:334] "Generic (PLEG): container finished" podID="19d490cc-8f49-4957-9812-94e96b24e100" containerID="4adc4530437a2232f9e63b1279dafeabf9af51275becdfd5424943bd1643b973" exitCode=0 Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.428014 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" event={"ID":"19d490cc-8f49-4957-9812-94e96b24e100","Type":"ContainerDied","Data":"4adc4530437a2232f9e63b1279dafeabf9af51275becdfd5424943bd1643b973"} Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.478830 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.489335 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-666b6646f7-64zgj" podUID="ccc57209-f9dc-4308-b11e-d3e10c70fe0a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.96:5353: connect: connection refused" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.489382 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.556541 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.574283 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19d490cc-8f49-4957-9812-94e96b24e100-dns-svc\") pod \"19d490cc-8f49-4957-9812-94e96b24e100\" (UID: \"19d490cc-8f49-4957-9812-94e96b24e100\") " Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.574383 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtn2l\" (UniqueName: \"kubernetes.io/projected/19d490cc-8f49-4957-9812-94e96b24e100-kube-api-access-qtn2l\") pod \"19d490cc-8f49-4957-9812-94e96b24e100\" (UID: \"19d490cc-8f49-4957-9812-94e96b24e100\") " Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.574439 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d490cc-8f49-4957-9812-94e96b24e100-config\") pod \"19d490cc-8f49-4957-9812-94e96b24e100\" (UID: \"19d490cc-8f49-4957-9812-94e96b24e100\") " Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.581242 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19d490cc-8f49-4957-9812-94e96b24e100-kube-api-access-qtn2l" (OuterVolumeSpecName: "kube-api-access-qtn2l") pod "19d490cc-8f49-4957-9812-94e96b24e100" (UID: "19d490cc-8f49-4957-9812-94e96b24e100"). InnerVolumeSpecName "kube-api-access-qtn2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.623455 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19d490cc-8f49-4957-9812-94e96b24e100-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "19d490cc-8f49-4957-9812-94e96b24e100" (UID: "19d490cc-8f49-4957-9812-94e96b24e100"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.651412 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19d490cc-8f49-4957-9812-94e96b24e100-config" (OuterVolumeSpecName: "config") pod "19d490cc-8f49-4957-9812-94e96b24e100" (UID: "19d490cc-8f49-4957-9812-94e96b24e100"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.676245 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19d490cc-8f49-4957-9812-94e96b24e100-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.676281 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtn2l\" (UniqueName: \"kubernetes.io/projected/19d490cc-8f49-4957-9812-94e96b24e100-kube-api-access-qtn2l\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.676293 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d490cc-8f49-4957-9812-94e96b24e100-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.776174 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-mgnmg"] Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.782716 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-tr4vq"] Oct 03 16:42:14 crc kubenswrapper[4744]: W1003 16:42:14.786823 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02bac763_b977_4d19_a6fd_300b0b2cefc3.slice/crio-597dfa57d0401b1a6e84630267070548584daa16363b9bdf9d0e694c1a73cb3e WatchSource:0}: Error finding container 597dfa57d0401b1a6e84630267070548584daa16363b9bdf9d0e694c1a73cb3e: Status 404 returned error can't find the container with id 597dfa57d0401b1a6e84630267070548584daa16363b9bdf9d0e694c1a73cb3e Oct 03 16:42:14 crc kubenswrapper[4744]: W1003 16:42:14.797009 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf8ea9a9_38eb_4588_82ef_831abd8841cc.slice/crio-33e40347e224c659b9bd163a946e01eab9621c692c6eaaf5e0038039428aae63 WatchSource:0}: Error finding container 33e40347e224c659b9bd163a946e01eab9621c692c6eaaf5e0038039428aae63: Status 404 returned error can't find the container with id 33e40347e224c659b9bd163a946e01eab9621c692c6eaaf5e0038039428aae63 Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.937042 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 03 16:42:14 crc kubenswrapper[4744]: E1003 16:42:14.937390 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d490cc-8f49-4957-9812-94e96b24e100" containerName="dnsmasq-dns" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.937405 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d490cc-8f49-4957-9812-94e96b24e100" containerName="dnsmasq-dns" Oct 03 16:42:14 crc kubenswrapper[4744]: E1003 16:42:14.937442 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d490cc-8f49-4957-9812-94e96b24e100" containerName="init" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.937449 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d490cc-8f49-4957-9812-94e96b24e100" containerName="init" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.937616 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="19d490cc-8f49-4957-9812-94e96b24e100" containerName="dnsmasq-dns" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.941313 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.941522 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.944441 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.944657 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.944811 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-6dxxg" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.944931 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.988549 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09af7ba6-34b5-4454-9dae-5bcc301ed20c-config\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.988603 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/09af7ba6-34b5-4454-9dae-5bcc301ed20c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.988637 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrl9r\" (UniqueName: \"kubernetes.io/projected/09af7ba6-34b5-4454-9dae-5bcc301ed20c-kube-api-access-nrl9r\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.988743 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/09af7ba6-34b5-4454-9dae-5bcc301ed20c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.988790 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09af7ba6-34b5-4454-9dae-5bcc301ed20c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.989236 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09af7ba6-34b5-4454-9dae-5bcc301ed20c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:14 crc kubenswrapper[4744]: I1003 16:42:14.990165 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09af7ba6-34b5-4454-9dae-5bcc301ed20c-scripts\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.022380 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9qqd6"] Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.091571 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09af7ba6-34b5-4454-9dae-5bcc301ed20c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.091620 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09af7ba6-34b5-4454-9dae-5bcc301ed20c-scripts\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.091658 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09af7ba6-34b5-4454-9dae-5bcc301ed20c-config\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.091676 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/09af7ba6-34b5-4454-9dae-5bcc301ed20c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.091696 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrl9r\" (UniqueName: \"kubernetes.io/projected/09af7ba6-34b5-4454-9dae-5bcc301ed20c-kube-api-access-nrl9r\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.091739 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/09af7ba6-34b5-4454-9dae-5bcc301ed20c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.091768 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09af7ba6-34b5-4454-9dae-5bcc301ed20c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.094062 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09af7ba6-34b5-4454-9dae-5bcc301ed20c-config\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.094792 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09af7ba6-34b5-4454-9dae-5bcc301ed20c-scripts\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.094955 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/09af7ba6-34b5-4454-9dae-5bcc301ed20c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.098896 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/09af7ba6-34b5-4454-9dae-5bcc301ed20c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.099341 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09af7ba6-34b5-4454-9dae-5bcc301ed20c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.101439 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09af7ba6-34b5-4454-9dae-5bcc301ed20c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.119467 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrl9r\" (UniqueName: \"kubernetes.io/projected/09af7ba6-34b5-4454-9dae-5bcc301ed20c-kube-api-access-nrl9r\") pod \"ovn-northd-0\" (UID: \"09af7ba6-34b5-4454-9dae-5bcc301ed20c\") " pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.206068 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-64zgj" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.294228 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg28q\" (UniqueName: \"kubernetes.io/projected/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-kube-api-access-wg28q\") pod \"ccc57209-f9dc-4308-b11e-d3e10c70fe0a\" (UID: \"ccc57209-f9dc-4308-b11e-d3e10c70fe0a\") " Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.294416 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-dns-svc\") pod \"ccc57209-f9dc-4308-b11e-d3e10c70fe0a\" (UID: \"ccc57209-f9dc-4308-b11e-d3e10c70fe0a\") " Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.294455 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-config\") pod \"ccc57209-f9dc-4308-b11e-d3e10c70fe0a\" (UID: \"ccc57209-f9dc-4308-b11e-d3e10c70fe0a\") " Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.302916 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-kube-api-access-wg28q" (OuterVolumeSpecName: "kube-api-access-wg28q") pod "ccc57209-f9dc-4308-b11e-d3e10c70fe0a" (UID: "ccc57209-f9dc-4308-b11e-d3e10c70fe0a"). InnerVolumeSpecName "kube-api-access-wg28q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.334025 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ccc57209-f9dc-4308-b11e-d3e10c70fe0a" (UID: "ccc57209-f9dc-4308-b11e-d3e10c70fe0a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.336066 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-config" (OuterVolumeSpecName: "config") pod "ccc57209-f9dc-4308-b11e-d3e10c70fe0a" (UID: "ccc57209-f9dc-4308-b11e-d3e10c70fe0a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.384837 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.396018 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.396041 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.396051 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg28q\" (UniqueName: \"kubernetes.io/projected/ccc57209-f9dc-4308-b11e-d3e10c70fe0a-kube-api-access-wg28q\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.454308 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.454338 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bwpkl" event={"ID":"19d490cc-8f49-4957-9812-94e96b24e100","Type":"ContainerDied","Data":"3d05c868402777cc598a7982adfa17791e6feaa03e807a7cb7049d72b6af3414"} Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.454427 4744 scope.go:117] "RemoveContainer" containerID="4adc4530437a2232f9e63b1279dafeabf9af51275becdfd5424943bd1643b973" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.462805 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-mgnmg" event={"ID":"02bac763-b977-4d19-a6fd-300b0b2cefc3","Type":"ContainerStarted","Data":"225ea9f0bb19a892557fac396424cd260113c3a3eeeb448b6c8dff78f6acb502"} Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.462837 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-mgnmg" event={"ID":"02bac763-b977-4d19-a6fd-300b0b2cefc3","Type":"ContainerStarted","Data":"597dfa57d0401b1a6e84630267070548584daa16363b9bdf9d0e694c1a73cb3e"} Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.468740 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-64zgj" event={"ID":"ccc57209-f9dc-4308-b11e-d3e10c70fe0a","Type":"ContainerDied","Data":"4613b61d2c7139f9bf97adae63885d82b2903577a043ecb231c44276b0c58e97"} Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.469177 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-64zgj" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.472058 4744 generic.go:334] "Generic (PLEG): container finished" podID="39bde481-e439-403f-8b03-8775c82ced64" containerID="69fbcdfb3c433265aa86ba5a6b4548d50dfe60f8f48e31682ee2ae226300973d" exitCode=0 Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.472171 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" event={"ID":"39bde481-e439-403f-8b03-8775c82ced64","Type":"ContainerDied","Data":"69fbcdfb3c433265aa86ba5a6b4548d50dfe60f8f48e31682ee2ae226300973d"} Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.472203 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" event={"ID":"39bde481-e439-403f-8b03-8775c82ced64","Type":"ContainerStarted","Data":"1c4844553f4f0b55aa3744e38bb6965d9813f46179a25699a39a6d0f3494bc64"} Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.476629 4744 generic.go:334] "Generic (PLEG): container finished" podID="cf8ea9a9-38eb-4588-82ef-831abd8841cc" containerID="77470324b69b64b6eeda0cfef2aec78ceb8716b56f4bd4f0048731b30146be52" exitCode=0 Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.478247 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" event={"ID":"cf8ea9a9-38eb-4588-82ef-831abd8841cc","Type":"ContainerDied","Data":"77470324b69b64b6eeda0cfef2aec78ceb8716b56f4bd4f0048731b30146be52"} Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.479404 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" event={"ID":"cf8ea9a9-38eb-4588-82ef-831abd8841cc","Type":"ContainerStarted","Data":"33e40347e224c659b9bd163a946e01eab9621c692c6eaaf5e0038039428aae63"} Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.482921 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bwpkl"] Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.508165 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-mgnmg" podStartSLOduration=2.50814469 podStartE2EDuration="2.50814469s" podCreationTimestamp="2025-10-03 16:42:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:42:15.490546666 +0000 UTC m=+1061.770422562" watchObservedRunningTime="2025-10-03 16:42:15.50814469 +0000 UTC m=+1061.788020586" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.508847 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bwpkl"] Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.517764 4744 scope.go:117] "RemoveContainer" containerID="4f1d7cbce0d835fa92818618aa79a0fd1f46830c20db70bea380e4c2cc675c65" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.579386 4744 scope.go:117] "RemoveContainer" containerID="2e5acad8a02a926b49ce792a513128c48289fccaeb3b94d4950d6370278f4373" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.587127 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-64zgj"] Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.596235 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-64zgj"] Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.613387 4744 scope.go:117] "RemoveContainer" containerID="bc3cc215bec8054bff3ce779162950ce5320408680bf2ba3e8a96355d9b6785e" Oct 03 16:42:15 crc kubenswrapper[4744]: I1003 16:42:15.929321 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 03 16:42:15 crc kubenswrapper[4744]: W1003 16:42:15.935721 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09af7ba6_34b5_4454_9dae_5bcc301ed20c.slice/crio-cf9b5500a2b8f6c7515f815017eb941f6e10dffd907a2dc9fe2989b2135d7d03 WatchSource:0}: Error finding container cf9b5500a2b8f6c7515f815017eb941f6e10dffd907a2dc9fe2989b2135d7d03: Status 404 returned error can't find the container with id cf9b5500a2b8f6c7515f815017eb941f6e10dffd907a2dc9fe2989b2135d7d03 Oct 03 16:42:16 crc kubenswrapper[4744]: I1003 16:42:16.488113 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" event={"ID":"cf8ea9a9-38eb-4588-82ef-831abd8841cc","Type":"ContainerStarted","Data":"8c701dd0ae3cabb9e52754b7237761e5a4ebd8f969934a541f0d51ca74725a39"} Oct 03 16:42:16 crc kubenswrapper[4744]: I1003 16:42:16.488568 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:16 crc kubenswrapper[4744]: I1003 16:42:16.493665 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"09af7ba6-34b5-4454-9dae-5bcc301ed20c","Type":"ContainerStarted","Data":"cf9b5500a2b8f6c7515f815017eb941f6e10dffd907a2dc9fe2989b2135d7d03"} Oct 03 16:42:16 crc kubenswrapper[4744]: I1003 16:42:16.495740 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" event={"ID":"39bde481-e439-403f-8b03-8775c82ced64","Type":"ContainerStarted","Data":"80b25fe51f20150897eb4859b590f8b41621ef08005004e33fcf8c60f75edd95"} Oct 03 16:42:16 crc kubenswrapper[4744]: I1003 16:42:16.511936 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" podStartSLOduration=3.511911293 podStartE2EDuration="3.511911293s" podCreationTimestamp="2025-10-03 16:42:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:42:16.509208494 +0000 UTC m=+1062.789084410" watchObservedRunningTime="2025-10-03 16:42:16.511911293 +0000 UTC m=+1062.791787199" Oct 03 16:42:16 crc kubenswrapper[4744]: I1003 16:42:16.538340 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" podStartSLOduration=2.538319845 podStartE2EDuration="2.538319845s" podCreationTimestamp="2025-10-03 16:42:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:42:16.535901483 +0000 UTC m=+1062.815777389" watchObservedRunningTime="2025-10-03 16:42:16.538319845 +0000 UTC m=+1062.818195751" Oct 03 16:42:16 crc kubenswrapper[4744]: I1003 16:42:16.909601 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19d490cc-8f49-4957-9812-94e96b24e100" path="/var/lib/kubelet/pods/19d490cc-8f49-4957-9812-94e96b24e100/volumes" Oct 03 16:42:16 crc kubenswrapper[4744]: I1003 16:42:16.910583 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccc57209-f9dc-4308-b11e-d3e10c70fe0a" path="/var/lib/kubelet/pods/ccc57209-f9dc-4308-b11e-d3e10c70fe0a/volumes" Oct 03 16:42:17 crc kubenswrapper[4744]: I1003 16:42:17.505927 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"09af7ba6-34b5-4454-9dae-5bcc301ed20c","Type":"ContainerStarted","Data":"33d5e6fbf4a2c586564bb6d509b546098521bfd4fb705c8f5863203967a4d338"} Oct 03 16:42:17 crc kubenswrapper[4744]: I1003 16:42:17.506468 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:18 crc kubenswrapper[4744]: I1003 16:42:18.060677 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 03 16:42:18 crc kubenswrapper[4744]: I1003 16:42:18.060791 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 03 16:42:18 crc kubenswrapper[4744]: I1003 16:42:18.140887 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 03 16:42:18 crc kubenswrapper[4744]: I1003 16:42:18.404168 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 03 16:42:18 crc kubenswrapper[4744]: I1003 16:42:18.404255 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 03 16:42:18 crc kubenswrapper[4744]: I1003 16:42:18.517583 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"09af7ba6-34b5-4454-9dae-5bcc301ed20c","Type":"ContainerStarted","Data":"a7d526915f490b8572bb7ab15e4038d096a717d64e73e54af64f5571c1e04d7c"} Oct 03 16:42:18 crc kubenswrapper[4744]: I1003 16:42:18.548909 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.3360061180000002 podStartE2EDuration="4.548881679s" podCreationTimestamp="2025-10-03 16:42:14 +0000 UTC" firstStartedPulling="2025-10-03 16:42:15.940841091 +0000 UTC m=+1062.220716987" lastFinishedPulling="2025-10-03 16:42:17.153716652 +0000 UTC m=+1063.433592548" observedRunningTime="2025-10-03 16:42:18.541464217 +0000 UTC m=+1064.821340123" watchObservedRunningTime="2025-10-03 16:42:18.548881679 +0000 UTC m=+1064.828757585" Oct 03 16:42:18 crc kubenswrapper[4744]: I1003 16:42:18.590277 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 03 16:42:18 crc kubenswrapper[4744]: I1003 16:42:18.832909 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-mj4kq"] Oct 03 16:42:18 crc kubenswrapper[4744]: E1003 16:42:18.833468 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc57209-f9dc-4308-b11e-d3e10c70fe0a" containerName="init" Oct 03 16:42:18 crc kubenswrapper[4744]: I1003 16:42:18.833550 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc57209-f9dc-4308-b11e-d3e10c70fe0a" containerName="init" Oct 03 16:42:18 crc kubenswrapper[4744]: E1003 16:42:18.833616 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc57209-f9dc-4308-b11e-d3e10c70fe0a" containerName="dnsmasq-dns" Oct 03 16:42:18 crc kubenswrapper[4744]: I1003 16:42:18.833677 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc57209-f9dc-4308-b11e-d3e10c70fe0a" containerName="dnsmasq-dns" Oct 03 16:42:18 crc kubenswrapper[4744]: I1003 16:42:18.833869 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccc57209-f9dc-4308-b11e-d3e10c70fe0a" containerName="dnsmasq-dns" Oct 03 16:42:18 crc kubenswrapper[4744]: I1003 16:42:18.834462 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mj4kq" Oct 03 16:42:18 crc kubenswrapper[4744]: I1003 16:42:18.844789 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-mj4kq"] Oct 03 16:42:18 crc kubenswrapper[4744]: I1003 16:42:18.966616 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h6lh\" (UniqueName: \"kubernetes.io/projected/b6f99e55-13bd-4d45-8ab3-2e4c06485dbe-kube-api-access-4h6lh\") pod \"placement-db-create-mj4kq\" (UID: \"b6f99e55-13bd-4d45-8ab3-2e4c06485dbe\") " pod="openstack/placement-db-create-mj4kq" Oct 03 16:42:19 crc kubenswrapper[4744]: I1003 16:42:19.069095 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h6lh\" (UniqueName: \"kubernetes.io/projected/b6f99e55-13bd-4d45-8ab3-2e4c06485dbe-kube-api-access-4h6lh\") pod \"placement-db-create-mj4kq\" (UID: \"b6f99e55-13bd-4d45-8ab3-2e4c06485dbe\") " pod="openstack/placement-db-create-mj4kq" Oct 03 16:42:19 crc kubenswrapper[4744]: I1003 16:42:19.094093 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h6lh\" (UniqueName: \"kubernetes.io/projected/b6f99e55-13bd-4d45-8ab3-2e4c06485dbe-kube-api-access-4h6lh\") pod \"placement-db-create-mj4kq\" (UID: \"b6f99e55-13bd-4d45-8ab3-2e4c06485dbe\") " pod="openstack/placement-db-create-mj4kq" Oct 03 16:42:19 crc kubenswrapper[4744]: I1003 16:42:19.201550 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mj4kq" Oct 03 16:42:19 crc kubenswrapper[4744]: I1003 16:42:19.526828 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 03 16:42:19 crc kubenswrapper[4744]: I1003 16:42:19.675929 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-mj4kq"] Oct 03 16:42:19 crc kubenswrapper[4744]: W1003 16:42:19.677999 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6f99e55_13bd_4d45_8ab3_2e4c06485dbe.slice/crio-e20e9b0267b4c06c41ebc390eae239874caf09017b87e065b06fcdc93b7a42dc WatchSource:0}: Error finding container e20e9b0267b4c06c41ebc390eae239874caf09017b87e065b06fcdc93b7a42dc: Status 404 returned error can't find the container with id e20e9b0267b4c06c41ebc390eae239874caf09017b87e065b06fcdc93b7a42dc Oct 03 16:42:20 crc kubenswrapper[4744]: I1003 16:42:20.487969 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 03 16:42:20 crc kubenswrapper[4744]: I1003 16:42:20.545776 4744 generic.go:334] "Generic (PLEG): container finished" podID="b6f99e55-13bd-4d45-8ab3-2e4c06485dbe" containerID="223036cb094324614faa919b4300eafff5e384637710226c1ca30e1ef4854d2e" exitCode=0 Oct 03 16:42:20 crc kubenswrapper[4744]: I1003 16:42:20.548111 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mj4kq" event={"ID":"b6f99e55-13bd-4d45-8ab3-2e4c06485dbe","Type":"ContainerDied","Data":"223036cb094324614faa919b4300eafff5e384637710226c1ca30e1ef4854d2e"} Oct 03 16:42:20 crc kubenswrapper[4744]: I1003 16:42:20.548185 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mj4kq" event={"ID":"b6f99e55-13bd-4d45-8ab3-2e4c06485dbe","Type":"ContainerStarted","Data":"e20e9b0267b4c06c41ebc390eae239874caf09017b87e065b06fcdc93b7a42dc"} Oct 03 16:42:20 crc kubenswrapper[4744]: I1003 16:42:20.582290 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 03 16:42:20 crc kubenswrapper[4744]: I1003 16:42:20.780139 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 03 16:42:20 crc kubenswrapper[4744]: I1003 16:42:20.904330 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-tr4vq"] Oct 03 16:42:20 crc kubenswrapper[4744]: I1003 16:42:20.904677 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" podUID="cf8ea9a9-38eb-4588-82ef-831abd8841cc" containerName="dnsmasq-dns" containerID="cri-o://8c701dd0ae3cabb9e52754b7237761e5a4ebd8f969934a541f0d51ca74725a39" gracePeriod=10 Oct 03 16:42:20 crc kubenswrapper[4744]: I1003 16:42:20.907298 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:20 crc kubenswrapper[4744]: I1003 16:42:20.928196 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-g6wjt"] Oct 03 16:42:20 crc kubenswrapper[4744]: I1003 16:42:20.929709 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:20 crc kubenswrapper[4744]: I1003 16:42:20.967352 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-g6wjt"] Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.005175 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-g6wjt\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.005226 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-g6wjt\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.005248 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-dns-svc\") pod \"dnsmasq-dns-698758b865-g6wjt\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.005318 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-config\") pod \"dnsmasq-dns-698758b865-g6wjt\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.005342 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6xkl\" (UniqueName: \"kubernetes.io/projected/a01c5ef1-aeea-431e-a744-2be2b1a28acf-kube-api-access-d6xkl\") pod \"dnsmasq-dns-698758b865-g6wjt\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.106929 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-config\") pod \"dnsmasq-dns-698758b865-g6wjt\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.107208 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6xkl\" (UniqueName: \"kubernetes.io/projected/a01c5ef1-aeea-431e-a744-2be2b1a28acf-kube-api-access-d6xkl\") pod \"dnsmasq-dns-698758b865-g6wjt\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.107306 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-g6wjt\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.107354 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-g6wjt\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.107376 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-dns-svc\") pod \"dnsmasq-dns-698758b865-g6wjt\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.108018 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-config\") pod \"dnsmasq-dns-698758b865-g6wjt\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.109614 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-g6wjt\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.109676 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-g6wjt\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.109996 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-dns-svc\") pod \"dnsmasq-dns-698758b865-g6wjt\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.131545 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6xkl\" (UniqueName: \"kubernetes.io/projected/a01c5ef1-aeea-431e-a744-2be2b1a28acf-kube-api-access-d6xkl\") pod \"dnsmasq-dns-698758b865-g6wjt\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.379608 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.384621 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.513294 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-dns-svc\") pod \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.514136 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42wfv\" (UniqueName: \"kubernetes.io/projected/cf8ea9a9-38eb-4588-82ef-831abd8841cc-kube-api-access-42wfv\") pod \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.514193 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-ovsdbserver-nb\") pod \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.514260 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-config\") pod \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\" (UID: \"cf8ea9a9-38eb-4588-82ef-831abd8841cc\") " Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.523720 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf8ea9a9-38eb-4588-82ef-831abd8841cc-kube-api-access-42wfv" (OuterVolumeSpecName: "kube-api-access-42wfv") pod "cf8ea9a9-38eb-4588-82ef-831abd8841cc" (UID: "cf8ea9a9-38eb-4588-82ef-831abd8841cc"). InnerVolumeSpecName "kube-api-access-42wfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.559913 4744 generic.go:334] "Generic (PLEG): container finished" podID="cf8ea9a9-38eb-4588-82ef-831abd8841cc" containerID="8c701dd0ae3cabb9e52754b7237761e5a4ebd8f969934a541f0d51ca74725a39" exitCode=0 Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.560162 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.561154 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" event={"ID":"cf8ea9a9-38eb-4588-82ef-831abd8841cc","Type":"ContainerDied","Data":"8c701dd0ae3cabb9e52754b7237761e5a4ebd8f969934a541f0d51ca74725a39"} Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.561520 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-tr4vq" event={"ID":"cf8ea9a9-38eb-4588-82ef-831abd8841cc","Type":"ContainerDied","Data":"33e40347e224c659b9bd163a946e01eab9621c692c6eaaf5e0038039428aae63"} Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.561557 4744 scope.go:117] "RemoveContainer" containerID="8c701dd0ae3cabb9e52754b7237761e5a4ebd8f969934a541f0d51ca74725a39" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.568243 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-config" (OuterVolumeSpecName: "config") pod "cf8ea9a9-38eb-4588-82ef-831abd8841cc" (UID: "cf8ea9a9-38eb-4588-82ef-831abd8841cc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.573455 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cf8ea9a9-38eb-4588-82ef-831abd8841cc" (UID: "cf8ea9a9-38eb-4588-82ef-831abd8841cc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.577953 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cf8ea9a9-38eb-4588-82ef-831abd8841cc" (UID: "cf8ea9a9-38eb-4588-82ef-831abd8841cc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.604315 4744 scope.go:117] "RemoveContainer" containerID="77470324b69b64b6eeda0cfef2aec78ceb8716b56f4bd4f0048731b30146be52" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.616381 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.616405 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42wfv\" (UniqueName: \"kubernetes.io/projected/cf8ea9a9-38eb-4588-82ef-831abd8841cc-kube-api-access-42wfv\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.616416 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.616428 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf8ea9a9-38eb-4588-82ef-831abd8841cc-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.634827 4744 scope.go:117] "RemoveContainer" containerID="8c701dd0ae3cabb9e52754b7237761e5a4ebd8f969934a541f0d51ca74725a39" Oct 03 16:42:21 crc kubenswrapper[4744]: E1003 16:42:21.635422 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c701dd0ae3cabb9e52754b7237761e5a4ebd8f969934a541f0d51ca74725a39\": container with ID starting with 8c701dd0ae3cabb9e52754b7237761e5a4ebd8f969934a541f0d51ca74725a39 not found: ID does not exist" containerID="8c701dd0ae3cabb9e52754b7237761e5a4ebd8f969934a541f0d51ca74725a39" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.635459 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c701dd0ae3cabb9e52754b7237761e5a4ebd8f969934a541f0d51ca74725a39"} err="failed to get container status \"8c701dd0ae3cabb9e52754b7237761e5a4ebd8f969934a541f0d51ca74725a39\": rpc error: code = NotFound desc = could not find container \"8c701dd0ae3cabb9e52754b7237761e5a4ebd8f969934a541f0d51ca74725a39\": container with ID starting with 8c701dd0ae3cabb9e52754b7237761e5a4ebd8f969934a541f0d51ca74725a39 not found: ID does not exist" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.635481 4744 scope.go:117] "RemoveContainer" containerID="77470324b69b64b6eeda0cfef2aec78ceb8716b56f4bd4f0048731b30146be52" Oct 03 16:42:21 crc kubenswrapper[4744]: E1003 16:42:21.635779 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77470324b69b64b6eeda0cfef2aec78ceb8716b56f4bd4f0048731b30146be52\": container with ID starting with 77470324b69b64b6eeda0cfef2aec78ceb8716b56f4bd4f0048731b30146be52 not found: ID does not exist" containerID="77470324b69b64b6eeda0cfef2aec78ceb8716b56f4bd4f0048731b30146be52" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.635799 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77470324b69b64b6eeda0cfef2aec78ceb8716b56f4bd4f0048731b30146be52"} err="failed to get container status \"77470324b69b64b6eeda0cfef2aec78ceb8716b56f4bd4f0048731b30146be52\": rpc error: code = NotFound desc = could not find container \"77470324b69b64b6eeda0cfef2aec78ceb8716b56f4bd4f0048731b30146be52\": container with ID starting with 77470324b69b64b6eeda0cfef2aec78ceb8716b56f4bd4f0048731b30146be52 not found: ID does not exist" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.949592 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-tr4vq"] Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.960451 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mj4kq" Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.970553 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-g6wjt"] Oct 03 16:42:21 crc kubenswrapper[4744]: I1003 16:42:21.974956 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-tr4vq"] Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.033187 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h6lh\" (UniqueName: \"kubernetes.io/projected/b6f99e55-13bd-4d45-8ab3-2e4c06485dbe-kube-api-access-4h6lh\") pod \"b6f99e55-13bd-4d45-8ab3-2e4c06485dbe\" (UID: \"b6f99e55-13bd-4d45-8ab3-2e4c06485dbe\") " Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.071993 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6f99e55-13bd-4d45-8ab3-2e4c06485dbe-kube-api-access-4h6lh" (OuterVolumeSpecName: "kube-api-access-4h6lh") pod "b6f99e55-13bd-4d45-8ab3-2e4c06485dbe" (UID: "b6f99e55-13bd-4d45-8ab3-2e4c06485dbe"). InnerVolumeSpecName "kube-api-access-4h6lh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.084205 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 03 16:42:22 crc kubenswrapper[4744]: E1003 16:42:22.084659 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf8ea9a9-38eb-4588-82ef-831abd8841cc" containerName="dnsmasq-dns" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.084674 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf8ea9a9-38eb-4588-82ef-831abd8841cc" containerName="dnsmasq-dns" Oct 03 16:42:22 crc kubenswrapper[4744]: E1003 16:42:22.084698 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6f99e55-13bd-4d45-8ab3-2e4c06485dbe" containerName="mariadb-database-create" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.084705 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6f99e55-13bd-4d45-8ab3-2e4c06485dbe" containerName="mariadb-database-create" Oct 03 16:42:22 crc kubenswrapper[4744]: E1003 16:42:22.084719 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf8ea9a9-38eb-4588-82ef-831abd8841cc" containerName="init" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.084725 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf8ea9a9-38eb-4588-82ef-831abd8841cc" containerName="init" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.084910 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf8ea9a9-38eb-4588-82ef-831abd8841cc" containerName="dnsmasq-dns" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.084932 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6f99e55-13bd-4d45-8ab3-2e4c06485dbe" containerName="mariadb-database-create" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.092554 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.107194 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-vszls" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.108715 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.108762 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.108730 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.120033 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.135617 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h6lh\" (UniqueName: \"kubernetes.io/projected/b6f99e55-13bd-4d45-8ab3-2e4c06485dbe-kube-api-access-4h6lh\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.237861 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.237939 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/133725f5-8ac3-4f45-b6c3-f7aea913711d-lock\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.238001 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dc26\" (UniqueName: \"kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-kube-api-access-8dc26\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.238047 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.238171 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/133725f5-8ac3-4f45-b6c3-f7aea913711d-cache\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.339867 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.339982 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/133725f5-8ac3-4f45-b6c3-f7aea913711d-cache\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.340032 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.340057 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/133725f5-8ac3-4f45-b6c3-f7aea913711d-lock\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.340085 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dc26\" (UniqueName: \"kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-kube-api-access-8dc26\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: E1003 16:42:22.340095 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 03 16:42:22 crc kubenswrapper[4744]: E1003 16:42:22.340131 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 03 16:42:22 crc kubenswrapper[4744]: E1003 16:42:22.340191 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift podName:133725f5-8ac3-4f45-b6c3-f7aea913711d nodeName:}" failed. No retries permitted until 2025-10-03 16:42:22.840169365 +0000 UTC m=+1069.120045331 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift") pod "swift-storage-0" (UID: "133725f5-8ac3-4f45-b6c3-f7aea913711d") : configmap "swift-ring-files" not found Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.340514 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.340555 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/133725f5-8ac3-4f45-b6c3-f7aea913711d-cache\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.340844 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/133725f5-8ac3-4f45-b6c3-f7aea913711d-lock\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.365383 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dc26\" (UniqueName: \"kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-kube-api-access-8dc26\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.383900 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.549953 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-dsc89"] Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.551080 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.554548 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.555534 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.560331 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-dsc89"] Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.565180 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.578424 4744 generic.go:334] "Generic (PLEG): container finished" podID="a01c5ef1-aeea-431e-a744-2be2b1a28acf" containerID="32d987bd39c8be18ca5c70041dca3bfbf8a1419d23aa8140ee93822fd9b69312" exitCode=0 Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.578527 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-g6wjt" event={"ID":"a01c5ef1-aeea-431e-a744-2be2b1a28acf","Type":"ContainerDied","Data":"32d987bd39c8be18ca5c70041dca3bfbf8a1419d23aa8140ee93822fd9b69312"} Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.578554 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-g6wjt" event={"ID":"a01c5ef1-aeea-431e-a744-2be2b1a28acf","Type":"ContainerStarted","Data":"80ba143779351234eb3d494bbd7b2e43a39d3ee313d01f2fc545e8a8f86ccef0"} Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.582046 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mj4kq" event={"ID":"b6f99e55-13bd-4d45-8ab3-2e4c06485dbe","Type":"ContainerDied","Data":"e20e9b0267b4c06c41ebc390eae239874caf09017b87e065b06fcdc93b7a42dc"} Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.582106 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e20e9b0267b4c06c41ebc390eae239874caf09017b87e065b06fcdc93b7a42dc" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.582256 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mj4kq" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.644741 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-combined-ca-bundle\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.644982 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0b88692d-9c4d-4c67-b70f-921810ca71e1-etc-swift\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.645139 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b88692d-9c4d-4c67-b70f-921810ca71e1-scripts\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.645250 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-swiftconf\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.645330 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qssk2\" (UniqueName: \"kubernetes.io/projected/0b88692d-9c4d-4c67-b70f-921810ca71e1-kube-api-access-qssk2\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.645403 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0b88692d-9c4d-4c67-b70f-921810ca71e1-ring-data-devices\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.645603 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-dispersionconf\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.747053 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-dispersionconf\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.747123 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-combined-ca-bundle\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.747150 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0b88692d-9c4d-4c67-b70f-921810ca71e1-etc-swift\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.747192 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b88692d-9c4d-4c67-b70f-921810ca71e1-scripts\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.747242 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-swiftconf\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.747265 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qssk2\" (UniqueName: \"kubernetes.io/projected/0b88692d-9c4d-4c67-b70f-921810ca71e1-kube-api-access-qssk2\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.747280 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0b88692d-9c4d-4c67-b70f-921810ca71e1-ring-data-devices\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.748021 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0b88692d-9c4d-4c67-b70f-921810ca71e1-etc-swift\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.748048 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0b88692d-9c4d-4c67-b70f-921810ca71e1-ring-data-devices\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.748750 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b88692d-9c4d-4c67-b70f-921810ca71e1-scripts\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.751454 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-dispersionconf\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.752120 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-swiftconf\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.752576 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-combined-ca-bundle\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.763638 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qssk2\" (UniqueName: \"kubernetes.io/projected/0b88692d-9c4d-4c67-b70f-921810ca71e1-kube-api-access-qssk2\") pod \"swift-ring-rebalance-dsc89\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.849069 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:22 crc kubenswrapper[4744]: E1003 16:42:22.849234 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 03 16:42:22 crc kubenswrapper[4744]: E1003 16:42:22.849254 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 03 16:42:22 crc kubenswrapper[4744]: E1003 16:42:22.849309 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift podName:133725f5-8ac3-4f45-b6c3-f7aea913711d nodeName:}" failed. No retries permitted until 2025-10-03 16:42:23.849293958 +0000 UTC m=+1070.129169854 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift") pod "swift-storage-0" (UID: "133725f5-8ac3-4f45-b6c3-f7aea913711d") : configmap "swift-ring-files" not found Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.882002 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:22 crc kubenswrapper[4744]: I1003 16:42:22.903737 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf8ea9a9-38eb-4588-82ef-831abd8841cc" path="/var/lib/kubelet/pods/cf8ea9a9-38eb-4588-82ef-831abd8841cc/volumes" Oct 03 16:42:23 crc kubenswrapper[4744]: I1003 16:42:23.323627 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-dsc89"] Oct 03 16:42:23 crc kubenswrapper[4744]: W1003 16:42:23.330202 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b88692d_9c4d_4c67_b70f_921810ca71e1.slice/crio-fb9f004f61891a16bfcbe6db387c2f637beb2707233f1d107ce676c6214226a7 WatchSource:0}: Error finding container fb9f004f61891a16bfcbe6db387c2f637beb2707233f1d107ce676c6214226a7: Status 404 returned error can't find the container with id fb9f004f61891a16bfcbe6db387c2f637beb2707233f1d107ce676c6214226a7 Oct 03 16:42:23 crc kubenswrapper[4744]: I1003 16:42:23.598979 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dsc89" event={"ID":"0b88692d-9c4d-4c67-b70f-921810ca71e1","Type":"ContainerStarted","Data":"fb9f004f61891a16bfcbe6db387c2f637beb2707233f1d107ce676c6214226a7"} Oct 03 16:42:23 crc kubenswrapper[4744]: I1003 16:42:23.601439 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-g6wjt" event={"ID":"a01c5ef1-aeea-431e-a744-2be2b1a28acf","Type":"ContainerStarted","Data":"8bc1ed4255353772b0f525b2c50735fff42a10b152105eb81c01847286ad05af"} Oct 03 16:42:23 crc kubenswrapper[4744]: I1003 16:42:23.601656 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:23 crc kubenswrapper[4744]: I1003 16:42:23.620648 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-g6wjt" podStartSLOduration=3.620629861 podStartE2EDuration="3.620629861s" podCreationTimestamp="2025-10-03 16:42:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:42:23.619275596 +0000 UTC m=+1069.899151502" watchObservedRunningTime="2025-10-03 16:42:23.620629861 +0000 UTC m=+1069.900505767" Oct 03 16:42:23 crc kubenswrapper[4744]: I1003 16:42:23.866014 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:23 crc kubenswrapper[4744]: E1003 16:42:23.866179 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 03 16:42:23 crc kubenswrapper[4744]: E1003 16:42:23.866192 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 03 16:42:23 crc kubenswrapper[4744]: E1003 16:42:23.866234 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift podName:133725f5-8ac3-4f45-b6c3-f7aea913711d nodeName:}" failed. No retries permitted until 2025-10-03 16:42:25.866220671 +0000 UTC m=+1072.146096567 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift") pod "swift-storage-0" (UID: "133725f5-8ac3-4f45-b6c3-f7aea913711d") : configmap "swift-ring-files" not found Oct 03 16:42:24 crc kubenswrapper[4744]: I1003 16:42:24.052742 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-pmltl"] Oct 03 16:42:24 crc kubenswrapper[4744]: I1003 16:42:24.053784 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pmltl" Oct 03 16:42:24 crc kubenswrapper[4744]: I1003 16:42:24.081178 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pmltl"] Oct 03 16:42:24 crc kubenswrapper[4744]: I1003 16:42:24.172676 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-226h4\" (UniqueName: \"kubernetes.io/projected/ef2c213d-480e-4470-b69b-890209b5a6fb-kube-api-access-226h4\") pod \"glance-db-create-pmltl\" (UID: \"ef2c213d-480e-4470-b69b-890209b5a6fb\") " pod="openstack/glance-db-create-pmltl" Oct 03 16:42:24 crc kubenswrapper[4744]: I1003 16:42:24.274437 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-226h4\" (UniqueName: \"kubernetes.io/projected/ef2c213d-480e-4470-b69b-890209b5a6fb-kube-api-access-226h4\") pod \"glance-db-create-pmltl\" (UID: \"ef2c213d-480e-4470-b69b-890209b5a6fb\") " pod="openstack/glance-db-create-pmltl" Oct 03 16:42:24 crc kubenswrapper[4744]: I1003 16:42:24.296375 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-226h4\" (UniqueName: \"kubernetes.io/projected/ef2c213d-480e-4470-b69b-890209b5a6fb-kube-api-access-226h4\") pod \"glance-db-create-pmltl\" (UID: \"ef2c213d-480e-4470-b69b-890209b5a6fb\") " pod="openstack/glance-db-create-pmltl" Oct 03 16:42:24 crc kubenswrapper[4744]: I1003 16:42:24.374426 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pmltl" Oct 03 16:42:24 crc kubenswrapper[4744]: I1003 16:42:24.558823 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:24 crc kubenswrapper[4744]: I1003 16:42:24.807464 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pmltl"] Oct 03 16:42:25 crc kubenswrapper[4744]: I1003 16:42:25.903699 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:25 crc kubenswrapper[4744]: E1003 16:42:25.903935 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 03 16:42:25 crc kubenswrapper[4744]: E1003 16:42:25.905992 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 03 16:42:25 crc kubenswrapper[4744]: E1003 16:42:25.906116 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift podName:133725f5-8ac3-4f45-b6c3-f7aea913711d nodeName:}" failed. No retries permitted until 2025-10-03 16:42:29.906096142 +0000 UTC m=+1076.185972038 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift") pod "swift-storage-0" (UID: "133725f5-8ac3-4f45-b6c3-f7aea913711d") : configmap "swift-ring-files" not found Oct 03 16:42:26 crc kubenswrapper[4744]: W1003 16:42:26.584132 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef2c213d_480e_4470_b69b_890209b5a6fb.slice/crio-1004de0d109af93477a9d5c5c1bc44630dd7ac84af6c210da4978969a7e27cc4 WatchSource:0}: Error finding container 1004de0d109af93477a9d5c5c1bc44630dd7ac84af6c210da4978969a7e27cc4: Status 404 returned error can't find the container with id 1004de0d109af93477a9d5c5c1bc44630dd7ac84af6c210da4978969a7e27cc4 Oct 03 16:42:26 crc kubenswrapper[4744]: I1003 16:42:26.629325 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pmltl" event={"ID":"ef2c213d-480e-4470-b69b-890209b5a6fb","Type":"ContainerStarted","Data":"1004de0d109af93477a9d5c5c1bc44630dd7ac84af6c210da4978969a7e27cc4"} Oct 03 16:42:27 crc kubenswrapper[4744]: I1003 16:42:27.638436 4744 generic.go:334] "Generic (PLEG): container finished" podID="ef2c213d-480e-4470-b69b-890209b5a6fb" containerID="e78967dddcc3d21e07e0bd426dd342ffd1f6530eb46d46212a740ca2d5c0baab" exitCode=0 Oct 03 16:42:27 crc kubenswrapper[4744]: I1003 16:42:27.638535 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pmltl" event={"ID":"ef2c213d-480e-4470-b69b-890209b5a6fb","Type":"ContainerDied","Data":"e78967dddcc3d21e07e0bd426dd342ffd1f6530eb46d46212a740ca2d5c0baab"} Oct 03 16:42:27 crc kubenswrapper[4744]: I1003 16:42:27.640839 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dsc89" event={"ID":"0b88692d-9c4d-4c67-b70f-921810ca71e1","Type":"ContainerStarted","Data":"f194150ecad52ed14833b6947806c080a558ba65366ed5497e4251644626b2d0"} Oct 03 16:42:27 crc kubenswrapper[4744]: I1003 16:42:27.687934 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-dsc89" podStartSLOduration=2.3718140229999998 podStartE2EDuration="5.687897361s" podCreationTimestamp="2025-10-03 16:42:22 +0000 UTC" firstStartedPulling="2025-10-03 16:42:23.333915019 +0000 UTC m=+1069.613790915" lastFinishedPulling="2025-10-03 16:42:26.649998357 +0000 UTC m=+1072.929874253" observedRunningTime="2025-10-03 16:42:27.685446868 +0000 UTC m=+1073.965322764" watchObservedRunningTime="2025-10-03 16:42:27.687897361 +0000 UTC m=+1073.967773307" Oct 03 16:42:28 crc kubenswrapper[4744]: I1003 16:42:28.459316 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-x55rd"] Oct 03 16:42:28 crc kubenswrapper[4744]: I1003 16:42:28.461400 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-x55rd" Oct 03 16:42:28 crc kubenswrapper[4744]: I1003 16:42:28.471970 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-x55rd"] Oct 03 16:42:28 crc kubenswrapper[4744]: I1003 16:42:28.556997 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsrdg\" (UniqueName: \"kubernetes.io/projected/2f0066c0-388f-4281-83f4-1fe5b0907d1c-kube-api-access-hsrdg\") pod \"keystone-db-create-x55rd\" (UID: \"2f0066c0-388f-4281-83f4-1fe5b0907d1c\") " pod="openstack/keystone-db-create-x55rd" Oct 03 16:42:28 crc kubenswrapper[4744]: I1003 16:42:28.658791 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsrdg\" (UniqueName: \"kubernetes.io/projected/2f0066c0-388f-4281-83f4-1fe5b0907d1c-kube-api-access-hsrdg\") pod \"keystone-db-create-x55rd\" (UID: \"2f0066c0-388f-4281-83f4-1fe5b0907d1c\") " pod="openstack/keystone-db-create-x55rd" Oct 03 16:42:28 crc kubenswrapper[4744]: I1003 16:42:28.688228 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsrdg\" (UniqueName: \"kubernetes.io/projected/2f0066c0-388f-4281-83f4-1fe5b0907d1c-kube-api-access-hsrdg\") pod \"keystone-db-create-x55rd\" (UID: \"2f0066c0-388f-4281-83f4-1fe5b0907d1c\") " pod="openstack/keystone-db-create-x55rd" Oct 03 16:42:28 crc kubenswrapper[4744]: I1003 16:42:28.783361 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-x55rd" Oct 03 16:42:28 crc kubenswrapper[4744]: I1003 16:42:28.892975 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-cc39-account-create-c9g7l"] Oct 03 16:42:28 crc kubenswrapper[4744]: I1003 16:42:28.895771 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cc39-account-create-c9g7l" Oct 03 16:42:28 crc kubenswrapper[4744]: I1003 16:42:28.899794 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 03 16:42:28 crc kubenswrapper[4744]: I1003 16:42:28.916911 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-cc39-account-create-c9g7l"] Oct 03 16:42:28 crc kubenswrapper[4744]: I1003 16:42:28.965052 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmnwt\" (UniqueName: \"kubernetes.io/projected/0a08b81c-0ca0-422e-8323-6e8ba91d1a31-kube-api-access-zmnwt\") pod \"placement-cc39-account-create-c9g7l\" (UID: \"0a08b81c-0ca0-422e-8323-6e8ba91d1a31\") " pod="openstack/placement-cc39-account-create-c9g7l" Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.047008 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pmltl" Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.067282 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmnwt\" (UniqueName: \"kubernetes.io/projected/0a08b81c-0ca0-422e-8323-6e8ba91d1a31-kube-api-access-zmnwt\") pod \"placement-cc39-account-create-c9g7l\" (UID: \"0a08b81c-0ca0-422e-8323-6e8ba91d1a31\") " pod="openstack/placement-cc39-account-create-c9g7l" Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.085872 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmnwt\" (UniqueName: \"kubernetes.io/projected/0a08b81c-0ca0-422e-8323-6e8ba91d1a31-kube-api-access-zmnwt\") pod \"placement-cc39-account-create-c9g7l\" (UID: \"0a08b81c-0ca0-422e-8323-6e8ba91d1a31\") " pod="openstack/placement-cc39-account-create-c9g7l" Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.168929 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-226h4\" (UniqueName: \"kubernetes.io/projected/ef2c213d-480e-4470-b69b-890209b5a6fb-kube-api-access-226h4\") pod \"ef2c213d-480e-4470-b69b-890209b5a6fb\" (UID: \"ef2c213d-480e-4470-b69b-890209b5a6fb\") " Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.172420 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef2c213d-480e-4470-b69b-890209b5a6fb-kube-api-access-226h4" (OuterVolumeSpecName: "kube-api-access-226h4") pod "ef2c213d-480e-4470-b69b-890209b5a6fb" (UID: "ef2c213d-480e-4470-b69b-890209b5a6fb"). InnerVolumeSpecName "kube-api-access-226h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.237114 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cc39-account-create-c9g7l" Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.271788 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-226h4\" (UniqueName: \"kubernetes.io/projected/ef2c213d-480e-4470-b69b-890209b5a6fb-kube-api-access-226h4\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.310766 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-x55rd"] Oct 03 16:42:29 crc kubenswrapper[4744]: W1003 16:42:29.322693 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f0066c0_388f_4281_83f4_1fe5b0907d1c.slice/crio-7283c46050ca29063e28b969cefc20a77e6383f134d2ea5445093eb2ce2e2a89 WatchSource:0}: Error finding container 7283c46050ca29063e28b969cefc20a77e6383f134d2ea5445093eb2ce2e2a89: Status 404 returned error can't find the container with id 7283c46050ca29063e28b969cefc20a77e6383f134d2ea5445093eb2ce2e2a89 Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.523928 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-cc39-account-create-c9g7l"] Oct 03 16:42:29 crc kubenswrapper[4744]: W1003 16:42:29.529150 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a08b81c_0ca0_422e_8323_6e8ba91d1a31.slice/crio-45fb9ee52f734305cfd988555f218234618833410690b5aa9596074bf1594055 WatchSource:0}: Error finding container 45fb9ee52f734305cfd988555f218234618833410690b5aa9596074bf1594055: Status 404 returned error can't find the container with id 45fb9ee52f734305cfd988555f218234618833410690b5aa9596074bf1594055 Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.657829 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cc39-account-create-c9g7l" event={"ID":"0a08b81c-0ca0-422e-8323-6e8ba91d1a31","Type":"ContainerStarted","Data":"45fb9ee52f734305cfd988555f218234618833410690b5aa9596074bf1594055"} Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.662303 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pmltl" Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.662364 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pmltl" event={"ID":"ef2c213d-480e-4470-b69b-890209b5a6fb","Type":"ContainerDied","Data":"1004de0d109af93477a9d5c5c1bc44630dd7ac84af6c210da4978969a7e27cc4"} Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.662435 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1004de0d109af93477a9d5c5c1bc44630dd7ac84af6c210da4978969a7e27cc4" Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.664342 4744 generic.go:334] "Generic (PLEG): container finished" podID="2f0066c0-388f-4281-83f4-1fe5b0907d1c" containerID="97fc922c13c202a20cfce82ada09283fd345b0c4b6eb06d74184978d86796d28" exitCode=0 Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.664369 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-x55rd" event={"ID":"2f0066c0-388f-4281-83f4-1fe5b0907d1c","Type":"ContainerDied","Data":"97fc922c13c202a20cfce82ada09283fd345b0c4b6eb06d74184978d86796d28"} Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.664389 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-x55rd" event={"ID":"2f0066c0-388f-4281-83f4-1fe5b0907d1c","Type":"ContainerStarted","Data":"7283c46050ca29063e28b969cefc20a77e6383f134d2ea5445093eb2ce2e2a89"} Oct 03 16:42:29 crc kubenswrapper[4744]: I1003 16:42:29.982411 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:29 crc kubenswrapper[4744]: E1003 16:42:29.982740 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 03 16:42:29 crc kubenswrapper[4744]: E1003 16:42:29.982763 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 03 16:42:29 crc kubenswrapper[4744]: E1003 16:42:29.982814 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift podName:133725f5-8ac3-4f45-b6c3-f7aea913711d nodeName:}" failed. No retries permitted until 2025-10-03 16:42:37.982795986 +0000 UTC m=+1084.262671882 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift") pod "swift-storage-0" (UID: "133725f5-8ac3-4f45-b6c3-f7aea913711d") : configmap "swift-ring-files" not found Oct 03 16:42:30 crc kubenswrapper[4744]: I1003 16:42:30.448715 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 03 16:42:30 crc kubenswrapper[4744]: I1003 16:42:30.673604 4744 generic.go:334] "Generic (PLEG): container finished" podID="0a08b81c-0ca0-422e-8323-6e8ba91d1a31" containerID="7465f321030bfc2785ae6ebb64ef1c046057a3a288da59ea404f41de37e25aae" exitCode=0 Oct 03 16:42:30 crc kubenswrapper[4744]: I1003 16:42:30.674049 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cc39-account-create-c9g7l" event={"ID":"0a08b81c-0ca0-422e-8323-6e8ba91d1a31","Type":"ContainerDied","Data":"7465f321030bfc2785ae6ebb64ef1c046057a3a288da59ea404f41de37e25aae"} Oct 03 16:42:31 crc kubenswrapper[4744]: I1003 16:42:31.107966 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-x55rd" Oct 03 16:42:31 crc kubenswrapper[4744]: I1003 16:42:31.206799 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsrdg\" (UniqueName: \"kubernetes.io/projected/2f0066c0-388f-4281-83f4-1fe5b0907d1c-kube-api-access-hsrdg\") pod \"2f0066c0-388f-4281-83f4-1fe5b0907d1c\" (UID: \"2f0066c0-388f-4281-83f4-1fe5b0907d1c\") " Oct 03 16:42:31 crc kubenswrapper[4744]: I1003 16:42:31.216844 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f0066c0-388f-4281-83f4-1fe5b0907d1c-kube-api-access-hsrdg" (OuterVolumeSpecName: "kube-api-access-hsrdg") pod "2f0066c0-388f-4281-83f4-1fe5b0907d1c" (UID: "2f0066c0-388f-4281-83f4-1fe5b0907d1c"). InnerVolumeSpecName "kube-api-access-hsrdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:42:31 crc kubenswrapper[4744]: I1003 16:42:31.309172 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsrdg\" (UniqueName: \"kubernetes.io/projected/2f0066c0-388f-4281-83f4-1fe5b0907d1c-kube-api-access-hsrdg\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:31 crc kubenswrapper[4744]: I1003 16:42:31.386457 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:42:31 crc kubenswrapper[4744]: I1003 16:42:31.444417 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9qqd6"] Oct 03 16:42:31 crc kubenswrapper[4744]: I1003 16:42:31.444789 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" podUID="39bde481-e439-403f-8b03-8775c82ced64" containerName="dnsmasq-dns" containerID="cri-o://80b25fe51f20150897eb4859b590f8b41621ef08005004e33fcf8c60f75edd95" gracePeriod=10 Oct 03 16:42:31 crc kubenswrapper[4744]: I1003 16:42:31.688398 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-x55rd" Oct 03 16:42:31 crc kubenswrapper[4744]: I1003 16:42:31.688385 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-x55rd" event={"ID":"2f0066c0-388f-4281-83f4-1fe5b0907d1c","Type":"ContainerDied","Data":"7283c46050ca29063e28b969cefc20a77e6383f134d2ea5445093eb2ce2e2a89"} Oct 03 16:42:31 crc kubenswrapper[4744]: I1003 16:42:31.688921 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7283c46050ca29063e28b969cefc20a77e6383f134d2ea5445093eb2ce2e2a89" Oct 03 16:42:31 crc kubenswrapper[4744]: I1003 16:42:31.690491 4744 generic.go:334] "Generic (PLEG): container finished" podID="39bde481-e439-403f-8b03-8775c82ced64" containerID="80b25fe51f20150897eb4859b590f8b41621ef08005004e33fcf8c60f75edd95" exitCode=0 Oct 03 16:42:31 crc kubenswrapper[4744]: I1003 16:42:31.690708 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" event={"ID":"39bde481-e439-403f-8b03-8775c82ced64","Type":"ContainerDied","Data":"80b25fe51f20150897eb4859b590f8b41621ef08005004e33fcf8c60f75edd95"} Oct 03 16:42:31 crc kubenswrapper[4744]: I1003 16:42:31.880688 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:31 crc kubenswrapper[4744]: I1003 16:42:31.980566 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cc39-account-create-c9g7l" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.037710 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-config\") pod \"39bde481-e439-403f-8b03-8775c82ced64\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.037867 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-ovsdbserver-nb\") pod \"39bde481-e439-403f-8b03-8775c82ced64\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.037948 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8m2f7\" (UniqueName: \"kubernetes.io/projected/39bde481-e439-403f-8b03-8775c82ced64-kube-api-access-8m2f7\") pod \"39bde481-e439-403f-8b03-8775c82ced64\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.037977 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-ovsdbserver-sb\") pod \"39bde481-e439-403f-8b03-8775c82ced64\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.038043 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-dns-svc\") pod \"39bde481-e439-403f-8b03-8775c82ced64\" (UID: \"39bde481-e439-403f-8b03-8775c82ced64\") " Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.046753 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39bde481-e439-403f-8b03-8775c82ced64-kube-api-access-8m2f7" (OuterVolumeSpecName: "kube-api-access-8m2f7") pod "39bde481-e439-403f-8b03-8775c82ced64" (UID: "39bde481-e439-403f-8b03-8775c82ced64"). InnerVolumeSpecName "kube-api-access-8m2f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.086967 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "39bde481-e439-403f-8b03-8775c82ced64" (UID: "39bde481-e439-403f-8b03-8775c82ced64"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.101156 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-config" (OuterVolumeSpecName: "config") pod "39bde481-e439-403f-8b03-8775c82ced64" (UID: "39bde481-e439-403f-8b03-8775c82ced64"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.101841 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "39bde481-e439-403f-8b03-8775c82ced64" (UID: "39bde481-e439-403f-8b03-8775c82ced64"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.103222 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "39bde481-e439-403f-8b03-8775c82ced64" (UID: "39bde481-e439-403f-8b03-8775c82ced64"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.140185 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmnwt\" (UniqueName: \"kubernetes.io/projected/0a08b81c-0ca0-422e-8323-6e8ba91d1a31-kube-api-access-zmnwt\") pod \"0a08b81c-0ca0-422e-8323-6e8ba91d1a31\" (UID: \"0a08b81c-0ca0-422e-8323-6e8ba91d1a31\") " Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.141326 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.141478 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.141613 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8m2f7\" (UniqueName: \"kubernetes.io/projected/39bde481-e439-403f-8b03-8775c82ced64-kube-api-access-8m2f7\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.141716 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.141809 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39bde481-e439-403f-8b03-8775c82ced64-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.146819 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a08b81c-0ca0-422e-8323-6e8ba91d1a31-kube-api-access-zmnwt" (OuterVolumeSpecName: "kube-api-access-zmnwt") pod "0a08b81c-0ca0-422e-8323-6e8ba91d1a31" (UID: "0a08b81c-0ca0-422e-8323-6e8ba91d1a31"). InnerVolumeSpecName "kube-api-access-zmnwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.242954 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmnwt\" (UniqueName: \"kubernetes.io/projected/0a08b81c-0ca0-422e-8323-6e8ba91d1a31-kube-api-access-zmnwt\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.699864 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.699829 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-9qqd6" event={"ID":"39bde481-e439-403f-8b03-8775c82ced64","Type":"ContainerDied","Data":"1c4844553f4f0b55aa3744e38bb6965d9813f46179a25699a39a6d0f3494bc64"} Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.700912 4744 scope.go:117] "RemoveContainer" containerID="80b25fe51f20150897eb4859b590f8b41621ef08005004e33fcf8c60f75edd95" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.701184 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cc39-account-create-c9g7l" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.713747 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cc39-account-create-c9g7l" event={"ID":"0a08b81c-0ca0-422e-8323-6e8ba91d1a31","Type":"ContainerDied","Data":"45fb9ee52f734305cfd988555f218234618833410690b5aa9596074bf1594055"} Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.713801 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45fb9ee52f734305cfd988555f218234618833410690b5aa9596074bf1594055" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.723818 4744 scope.go:117] "RemoveContainer" containerID="69fbcdfb3c433265aa86ba5a6b4548d50dfe60f8f48e31682ee2ae226300973d" Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.738466 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9qqd6"] Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.755466 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9qqd6"] Oct 03 16:42:32 crc kubenswrapper[4744]: I1003 16:42:32.909146 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39bde481-e439-403f-8b03-8775c82ced64" path="/var/lib/kubelet/pods/39bde481-e439-403f-8b03-8775c82ced64/volumes" Oct 03 16:42:33 crc kubenswrapper[4744]: I1003 16:42:33.717294 4744 generic.go:334] "Generic (PLEG): container finished" podID="0b88692d-9c4d-4c67-b70f-921810ca71e1" containerID="f194150ecad52ed14833b6947806c080a558ba65366ed5497e4251644626b2d0" exitCode=0 Oct 03 16:42:33 crc kubenswrapper[4744]: I1003 16:42:33.717353 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dsc89" event={"ID":"0b88692d-9c4d-4c67-b70f-921810ca71e1","Type":"ContainerDied","Data":"f194150ecad52ed14833b6947806c080a558ba65366ed5497e4251644626b2d0"} Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.083669 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.196012 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-swiftconf\") pod \"0b88692d-9c4d-4c67-b70f-921810ca71e1\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.196064 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b88692d-9c4d-4c67-b70f-921810ca71e1-scripts\") pod \"0b88692d-9c4d-4c67-b70f-921810ca71e1\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.196113 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-dispersionconf\") pod \"0b88692d-9c4d-4c67-b70f-921810ca71e1\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.196232 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qssk2\" (UniqueName: \"kubernetes.io/projected/0b88692d-9c4d-4c67-b70f-921810ca71e1-kube-api-access-qssk2\") pod \"0b88692d-9c4d-4c67-b70f-921810ca71e1\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.196264 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0b88692d-9c4d-4c67-b70f-921810ca71e1-etc-swift\") pod \"0b88692d-9c4d-4c67-b70f-921810ca71e1\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.196290 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0b88692d-9c4d-4c67-b70f-921810ca71e1-ring-data-devices\") pod \"0b88692d-9c4d-4c67-b70f-921810ca71e1\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.196364 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-combined-ca-bundle\") pod \"0b88692d-9c4d-4c67-b70f-921810ca71e1\" (UID: \"0b88692d-9c4d-4c67-b70f-921810ca71e1\") " Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.201825 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b88692d-9c4d-4c67-b70f-921810ca71e1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "0b88692d-9c4d-4c67-b70f-921810ca71e1" (UID: "0b88692d-9c4d-4c67-b70f-921810ca71e1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.202267 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b88692d-9c4d-4c67-b70f-921810ca71e1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "0b88692d-9c4d-4c67-b70f-921810ca71e1" (UID: "0b88692d-9c4d-4c67-b70f-921810ca71e1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.204364 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b88692d-9c4d-4c67-b70f-921810ca71e1-kube-api-access-qssk2" (OuterVolumeSpecName: "kube-api-access-qssk2") pod "0b88692d-9c4d-4c67-b70f-921810ca71e1" (UID: "0b88692d-9c4d-4c67-b70f-921810ca71e1"). InnerVolumeSpecName "kube-api-access-qssk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.204847 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "0b88692d-9c4d-4c67-b70f-921810ca71e1" (UID: "0b88692d-9c4d-4c67-b70f-921810ca71e1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.217769 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b88692d-9c4d-4c67-b70f-921810ca71e1-scripts" (OuterVolumeSpecName: "scripts") pod "0b88692d-9c4d-4c67-b70f-921810ca71e1" (UID: "0b88692d-9c4d-4c67-b70f-921810ca71e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.220125 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "0b88692d-9c4d-4c67-b70f-921810ca71e1" (UID: "0b88692d-9c4d-4c67-b70f-921810ca71e1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.224796 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b88692d-9c4d-4c67-b70f-921810ca71e1" (UID: "0b88692d-9c4d-4c67-b70f-921810ca71e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.298420 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qssk2\" (UniqueName: \"kubernetes.io/projected/0b88692d-9c4d-4c67-b70f-921810ca71e1-kube-api-access-qssk2\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.298458 4744 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0b88692d-9c4d-4c67-b70f-921810ca71e1-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.298467 4744 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0b88692d-9c4d-4c67-b70f-921810ca71e1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.298476 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.298485 4744 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.298496 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b88692d-9c4d-4c67-b70f-921810ca71e1-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.298522 4744 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0b88692d-9c4d-4c67-b70f-921810ca71e1-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.740553 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-dsc89" event={"ID":"0b88692d-9c4d-4c67-b70f-921810ca71e1","Type":"ContainerDied","Data":"fb9f004f61891a16bfcbe6db387c2f637beb2707233f1d107ce676c6214226a7"} Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.740923 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb9f004f61891a16bfcbe6db387c2f637beb2707233f1d107ce676c6214226a7" Oct 03 16:42:35 crc kubenswrapper[4744]: I1003 16:42:35.740830 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dsc89" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.045161 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.057873 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/133725f5-8ac3-4f45-b6c3-f7aea913711d-etc-swift\") pod \"swift-storage-0\" (UID: \"133725f5-8ac3-4f45-b6c3-f7aea913711d\") " pod="openstack/swift-storage-0" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.326022 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.577340 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-d48a-account-create-t4tk2"] Oct 03 16:42:38 crc kubenswrapper[4744]: E1003 16:42:38.578102 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b88692d-9c4d-4c67-b70f-921810ca71e1" containerName="swift-ring-rebalance" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.578117 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b88692d-9c4d-4c67-b70f-921810ca71e1" containerName="swift-ring-rebalance" Oct 03 16:42:38 crc kubenswrapper[4744]: E1003 16:42:38.578130 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39bde481-e439-403f-8b03-8775c82ced64" containerName="init" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.578138 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="39bde481-e439-403f-8b03-8775c82ced64" containerName="init" Oct 03 16:42:38 crc kubenswrapper[4744]: E1003 16:42:38.578147 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2c213d-480e-4470-b69b-890209b5a6fb" containerName="mariadb-database-create" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.578155 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2c213d-480e-4470-b69b-890209b5a6fb" containerName="mariadb-database-create" Oct 03 16:42:38 crc kubenswrapper[4744]: E1003 16:42:38.578173 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a08b81c-0ca0-422e-8323-6e8ba91d1a31" containerName="mariadb-account-create" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.578181 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a08b81c-0ca0-422e-8323-6e8ba91d1a31" containerName="mariadb-account-create" Oct 03 16:42:38 crc kubenswrapper[4744]: E1003 16:42:38.578220 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0066c0-388f-4281-83f4-1fe5b0907d1c" containerName="mariadb-database-create" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.578230 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0066c0-388f-4281-83f4-1fe5b0907d1c" containerName="mariadb-database-create" Oct 03 16:42:38 crc kubenswrapper[4744]: E1003 16:42:38.578246 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39bde481-e439-403f-8b03-8775c82ced64" containerName="dnsmasq-dns" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.578254 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="39bde481-e439-403f-8b03-8775c82ced64" containerName="dnsmasq-dns" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.578462 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef2c213d-480e-4470-b69b-890209b5a6fb" containerName="mariadb-database-create" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.578476 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b88692d-9c4d-4c67-b70f-921810ca71e1" containerName="swift-ring-rebalance" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.578587 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="39bde481-e439-403f-8b03-8775c82ced64" containerName="dnsmasq-dns" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.578605 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a08b81c-0ca0-422e-8323-6e8ba91d1a31" containerName="mariadb-account-create" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.578615 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f0066c0-388f-4281-83f4-1fe5b0907d1c" containerName="mariadb-database-create" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.579602 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d48a-account-create-t4tk2" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.582024 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.593714 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d48a-account-create-t4tk2"] Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.657250 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t726x\" (UniqueName: \"kubernetes.io/projected/c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5-kube-api-access-t726x\") pod \"keystone-d48a-account-create-t4tk2\" (UID: \"c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5\") " pod="openstack/keystone-d48a-account-create-t4tk2" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.758612 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t726x\" (UniqueName: \"kubernetes.io/projected/c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5-kube-api-access-t726x\") pod \"keystone-d48a-account-create-t4tk2\" (UID: \"c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5\") " pod="openstack/keystone-d48a-account-create-t4tk2" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.779270 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t726x\" (UniqueName: \"kubernetes.io/projected/c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5-kube-api-access-t726x\") pod \"keystone-d48a-account-create-t4tk2\" (UID: \"c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5\") " pod="openstack/keystone-d48a-account-create-t4tk2" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.899068 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d48a-account-create-t4tk2" Oct 03 16:42:38 crc kubenswrapper[4744]: I1003 16:42:38.910057 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.329920 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d48a-account-create-t4tk2"] Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.562267 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4jn8x" podUID="bc8fb837-d1e5-4953-ae6c-fbee5f867e49" containerName="ovn-controller" probeResult="failure" output=< Oct 03 16:42:39 crc kubenswrapper[4744]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 03 16:42:39 crc kubenswrapper[4744]: > Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.626243 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.640302 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-99k4w" Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.771072 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"ce4184c5e31d391836eb91d5b7e5f169f7aee273d8997d95d06a3eb0c60ef574"} Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.772639 4744 generic.go:334] "Generic (PLEG): container finished" podID="c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5" containerID="3a939c13f0ada6e3771f4b1177d843e7975333c60e7bd85ce7a7a3e24e5ab528" exitCode=0 Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.772707 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d48a-account-create-t4tk2" event={"ID":"c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5","Type":"ContainerDied","Data":"3a939c13f0ada6e3771f4b1177d843e7975333c60e7bd85ce7a7a3e24e5ab528"} Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.772765 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d48a-account-create-t4tk2" event={"ID":"c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5","Type":"ContainerStarted","Data":"7dd3fb2085d29e223618e54aa258c8b7a865571ffe409006cff448cff87fe1b3"} Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.839975 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4jn8x-config-w5dh2"] Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.841191 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.848368 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.864112 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4jn8x-config-w5dh2"] Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.979106 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-run-ovn\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.979229 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/139159c5-0601-42e2-b71f-1497543df40f-scripts\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.979257 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-run\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.979276 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/139159c5-0601-42e2-b71f-1497543df40f-additional-scripts\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.979301 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-log-ovn\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:39 crc kubenswrapper[4744]: I1003 16:42:39.979320 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hvd7\" (UniqueName: \"kubernetes.io/projected/139159c5-0601-42e2-b71f-1497543df40f-kube-api-access-9hvd7\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:40 crc kubenswrapper[4744]: I1003 16:42:40.081096 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/139159c5-0601-42e2-b71f-1497543df40f-scripts\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:40 crc kubenswrapper[4744]: I1003 16:42:40.081432 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-run\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:40 crc kubenswrapper[4744]: I1003 16:42:40.081460 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/139159c5-0601-42e2-b71f-1497543df40f-additional-scripts\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:40 crc kubenswrapper[4744]: I1003 16:42:40.081489 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-log-ovn\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:40 crc kubenswrapper[4744]: I1003 16:42:40.081525 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hvd7\" (UniqueName: \"kubernetes.io/projected/139159c5-0601-42e2-b71f-1497543df40f-kube-api-access-9hvd7\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:40 crc kubenswrapper[4744]: I1003 16:42:40.081567 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-run-ovn\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:40 crc kubenswrapper[4744]: I1003 16:42:40.081763 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-run-ovn\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:40 crc kubenswrapper[4744]: I1003 16:42:40.081774 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-run\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:40 crc kubenswrapper[4744]: I1003 16:42:40.081814 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-log-ovn\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:40 crc kubenswrapper[4744]: I1003 16:42:40.082348 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/139159c5-0601-42e2-b71f-1497543df40f-additional-scripts\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:40 crc kubenswrapper[4744]: I1003 16:42:40.083058 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/139159c5-0601-42e2-b71f-1497543df40f-scripts\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:40 crc kubenswrapper[4744]: I1003 16:42:40.101658 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hvd7\" (UniqueName: \"kubernetes.io/projected/139159c5-0601-42e2-b71f-1497543df40f-kube-api-access-9hvd7\") pod \"ovn-controller-4jn8x-config-w5dh2\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:40 crc kubenswrapper[4744]: I1003 16:42:40.159677 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:40 crc kubenswrapper[4744]: I1003 16:42:40.806905 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"787cc0129e20f8472023dfc2b7713a40587aec2aeb4211f3bbd862489efe83d4"} Oct 03 16:42:40 crc kubenswrapper[4744]: I1003 16:42:40.992901 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4jn8x-config-w5dh2"] Oct 03 16:42:41 crc kubenswrapper[4744]: W1003 16:42:41.002999 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod139159c5_0601_42e2_b71f_1497543df40f.slice/crio-786c59f1e2319f7219a8e5a6c84d27ac5b0145487b977b11cb3011798b45e8bc WatchSource:0}: Error finding container 786c59f1e2319f7219a8e5a6c84d27ac5b0145487b977b11cb3011798b45e8bc: Status 404 returned error can't find the container with id 786c59f1e2319f7219a8e5a6c84d27ac5b0145487b977b11cb3011798b45e8bc Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.194015 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d48a-account-create-t4tk2" Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.304188 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t726x\" (UniqueName: \"kubernetes.io/projected/c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5-kube-api-access-t726x\") pod \"c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5\" (UID: \"c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5\") " Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.310789 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5-kube-api-access-t726x" (OuterVolumeSpecName: "kube-api-access-t726x") pod "c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5" (UID: "c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5"). InnerVolumeSpecName "kube-api-access-t726x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.405894 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t726x\" (UniqueName: \"kubernetes.io/projected/c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5-kube-api-access-t726x\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.820370 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"078b3525cca1c2b505606bf7db4562e92212d51beb5989f4349e9938ae50537e"} Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.820426 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"ccaca35fe94bba38af157a4d4c7f18af47ea51d2dacb486692066b8133c4afb3"} Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.820448 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"3f238088180ed83c498712f02fe4be3fd1b8c876d9084ead5b378eba46ca1f9d"} Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.822794 4744 generic.go:334] "Generic (PLEG): container finished" podID="7d21c39f-eece-4cf0-847b-c8ebc87c574d" containerID="50822061c1b75d267e0d3c6f99f1d88af5ebf5ff228151acf820e4a5ab291a81" exitCode=0 Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.822876 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d21c39f-eece-4cf0-847b-c8ebc87c574d","Type":"ContainerDied","Data":"50822061c1b75d267e0d3c6f99f1d88af5ebf5ff228151acf820e4a5ab291a81"} Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.829368 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d48a-account-create-t4tk2" event={"ID":"c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5","Type":"ContainerDied","Data":"7dd3fb2085d29e223618e54aa258c8b7a865571ffe409006cff448cff87fe1b3"} Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.829680 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dd3fb2085d29e223618e54aa258c8b7a865571ffe409006cff448cff87fe1b3" Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.829680 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d48a-account-create-t4tk2" Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.837969 4744 generic.go:334] "Generic (PLEG): container finished" podID="a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" containerID="468e9fb03dafde6384d03462697979e76f31467f870859d2320a59286b385cf5" exitCode=0 Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.838033 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533","Type":"ContainerDied","Data":"468e9fb03dafde6384d03462697979e76f31467f870859d2320a59286b385cf5"} Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.842663 4744 generic.go:334] "Generic (PLEG): container finished" podID="139159c5-0601-42e2-b71f-1497543df40f" containerID="1a78d69196e0287df60bf4c840b2a9b6309794774bad9c0eee264bc8bfa2c004" exitCode=0 Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.842706 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4jn8x-config-w5dh2" event={"ID":"139159c5-0601-42e2-b71f-1497543df40f","Type":"ContainerDied","Data":"1a78d69196e0287df60bf4c840b2a9b6309794774bad9c0eee264bc8bfa2c004"} Oct 03 16:42:41 crc kubenswrapper[4744]: I1003 16:42:41.842734 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4jn8x-config-w5dh2" event={"ID":"139159c5-0601-42e2-b71f-1497543df40f","Type":"ContainerStarted","Data":"786c59f1e2319f7219a8e5a6c84d27ac5b0145487b977b11cb3011798b45e8bc"} Oct 03 16:42:42 crc kubenswrapper[4744]: I1003 16:42:42.850832 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d21c39f-eece-4cf0-847b-c8ebc87c574d","Type":"ContainerStarted","Data":"355e3b5a1d962d3b2118f7e62c33c0d1c971a6aeb3586298b8b496211b6b4eb9"} Oct 03 16:42:42 crc kubenswrapper[4744]: I1003 16:42:42.851646 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 03 16:42:42 crc kubenswrapper[4744]: I1003 16:42:42.853429 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533","Type":"ContainerStarted","Data":"c8e8da6d287f224992ddd8696bd60c618abf8af3e621a9504b2b841c36ad6bca"} Oct 03 16:42:42 crc kubenswrapper[4744]: I1003 16:42:42.853856 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:42:42 crc kubenswrapper[4744]: I1003 16:42:42.880334 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=48.807076108 podStartE2EDuration="58.880314137s" podCreationTimestamp="2025-10-03 16:41:44 +0000 UTC" firstStartedPulling="2025-10-03 16:41:56.804697135 +0000 UTC m=+1043.084573031" lastFinishedPulling="2025-10-03 16:42:06.877935124 +0000 UTC m=+1053.157811060" observedRunningTime="2025-10-03 16:42:42.874747453 +0000 UTC m=+1089.154623359" watchObservedRunningTime="2025-10-03 16:42:42.880314137 +0000 UTC m=+1089.160190033" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.225421 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=49.998530995 podStartE2EDuration="1m0.225403951s" podCreationTimestamp="2025-10-03 16:41:44 +0000 UTC" firstStartedPulling="2025-10-03 16:41:56.848836384 +0000 UTC m=+1043.128712280" lastFinishedPulling="2025-10-03 16:42:07.07570933 +0000 UTC m=+1053.355585236" observedRunningTime="2025-10-03 16:42:42.898483696 +0000 UTC m=+1089.178359592" watchObservedRunningTime="2025-10-03 16:42:44.225403951 +0000 UTC m=+1090.505279847" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.229533 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-5696-account-create-jn6hh"] Oct 03 16:42:44 crc kubenswrapper[4744]: E1003 16:42:44.229871 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5" containerName="mariadb-account-create" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.229887 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5" containerName="mariadb-account-create" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.230060 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5" containerName="mariadb-account-create" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.230671 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5696-account-create-jn6hh" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.232282 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.246872 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-5696-account-create-jn6hh"] Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.248608 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.352757 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-run-ovn\") pod \"139159c5-0601-42e2-b71f-1497543df40f\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.352837 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "139159c5-0601-42e2-b71f-1497543df40f" (UID: "139159c5-0601-42e2-b71f-1497543df40f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.352871 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-run\") pod \"139159c5-0601-42e2-b71f-1497543df40f\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.352892 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-run" (OuterVolumeSpecName: "var-run") pod "139159c5-0601-42e2-b71f-1497543df40f" (UID: "139159c5-0601-42e2-b71f-1497543df40f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.352919 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/139159c5-0601-42e2-b71f-1497543df40f-additional-scripts\") pod \"139159c5-0601-42e2-b71f-1497543df40f\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.353012 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-log-ovn\") pod \"139159c5-0601-42e2-b71f-1497543df40f\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.353145 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "139159c5-0601-42e2-b71f-1497543df40f" (UID: "139159c5-0601-42e2-b71f-1497543df40f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.353158 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/139159c5-0601-42e2-b71f-1497543df40f-scripts\") pod \"139159c5-0601-42e2-b71f-1497543df40f\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.353239 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hvd7\" (UniqueName: \"kubernetes.io/projected/139159c5-0601-42e2-b71f-1497543df40f-kube-api-access-9hvd7\") pod \"139159c5-0601-42e2-b71f-1497543df40f\" (UID: \"139159c5-0601-42e2-b71f-1497543df40f\") " Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.353715 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/139159c5-0601-42e2-b71f-1497543df40f-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "139159c5-0601-42e2-b71f-1497543df40f" (UID: "139159c5-0601-42e2-b71f-1497543df40f"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.353753 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsj2w\" (UniqueName: \"kubernetes.io/projected/d34f7f18-2333-4975-9606-9dd4ec282f3e-kube-api-access-nsj2w\") pod \"glance-5696-account-create-jn6hh\" (UID: \"d34f7f18-2333-4975-9606-9dd4ec282f3e\") " pod="openstack/glance-5696-account-create-jn6hh" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.353976 4744 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.353995 4744 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-run\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.354006 4744 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/139159c5-0601-42e2-b71f-1497543df40f-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.354016 4744 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/139159c5-0601-42e2-b71f-1497543df40f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.354881 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/139159c5-0601-42e2-b71f-1497543df40f-scripts" (OuterVolumeSpecName: "scripts") pod "139159c5-0601-42e2-b71f-1497543df40f" (UID: "139159c5-0601-42e2-b71f-1497543df40f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.365875 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/139159c5-0601-42e2-b71f-1497543df40f-kube-api-access-9hvd7" (OuterVolumeSpecName: "kube-api-access-9hvd7") pod "139159c5-0601-42e2-b71f-1497543df40f" (UID: "139159c5-0601-42e2-b71f-1497543df40f"). InnerVolumeSpecName "kube-api-access-9hvd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.455426 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsj2w\" (UniqueName: \"kubernetes.io/projected/d34f7f18-2333-4975-9606-9dd4ec282f3e-kube-api-access-nsj2w\") pod \"glance-5696-account-create-jn6hh\" (UID: \"d34f7f18-2333-4975-9606-9dd4ec282f3e\") " pod="openstack/glance-5696-account-create-jn6hh" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.455547 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/139159c5-0601-42e2-b71f-1497543df40f-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.455560 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hvd7\" (UniqueName: \"kubernetes.io/projected/139159c5-0601-42e2-b71f-1497543df40f-kube-api-access-9hvd7\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.476587 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsj2w\" (UniqueName: \"kubernetes.io/projected/d34f7f18-2333-4975-9606-9dd4ec282f3e-kube-api-access-nsj2w\") pod \"glance-5696-account-create-jn6hh\" (UID: \"d34f7f18-2333-4975-9606-9dd4ec282f3e\") " pod="openstack/glance-5696-account-create-jn6hh" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.559274 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5696-account-create-jn6hh" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.608814 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-4jn8x" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.873903 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4jn8x-config-w5dh2" event={"ID":"139159c5-0601-42e2-b71f-1497543df40f","Type":"ContainerDied","Data":"786c59f1e2319f7219a8e5a6c84d27ac5b0145487b977b11cb3011798b45e8bc"} Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.874225 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="786c59f1e2319f7219a8e5a6c84d27ac5b0145487b977b11cb3011798b45e8bc" Oct 03 16:42:44 crc kubenswrapper[4744]: I1003 16:42:44.873963 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4jn8x-config-w5dh2" Oct 03 16:42:45 crc kubenswrapper[4744]: I1003 16:42:45.037381 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-5696-account-create-jn6hh"] Oct 03 16:42:45 crc kubenswrapper[4744]: W1003 16:42:45.234125 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd34f7f18_2333_4975_9606_9dd4ec282f3e.slice/crio-1ed1a2167332877b03c55ee5b2215fb47721327084ac99b337b241158b70202b WatchSource:0}: Error finding container 1ed1a2167332877b03c55ee5b2215fb47721327084ac99b337b241158b70202b: Status 404 returned error can't find the container with id 1ed1a2167332877b03c55ee5b2215fb47721327084ac99b337b241158b70202b Oct 03 16:42:45 crc kubenswrapper[4744]: I1003 16:42:45.361625 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4jn8x-config-w5dh2"] Oct 03 16:42:45 crc kubenswrapper[4744]: I1003 16:42:45.367540 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4jn8x-config-w5dh2"] Oct 03 16:42:45 crc kubenswrapper[4744]: I1003 16:42:45.888690 4744 generic.go:334] "Generic (PLEG): container finished" podID="d34f7f18-2333-4975-9606-9dd4ec282f3e" containerID="d9506b1991e30e193a66b300a91f395b1c2458fa895fdd1f8798c4753a65a1ba" exitCode=0 Oct 03 16:42:45 crc kubenswrapper[4744]: I1003 16:42:45.888930 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5696-account-create-jn6hh" event={"ID":"d34f7f18-2333-4975-9606-9dd4ec282f3e","Type":"ContainerDied","Data":"d9506b1991e30e193a66b300a91f395b1c2458fa895fdd1f8798c4753a65a1ba"} Oct 03 16:42:45 crc kubenswrapper[4744]: I1003 16:42:45.889148 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5696-account-create-jn6hh" event={"ID":"d34f7f18-2333-4975-9606-9dd4ec282f3e","Type":"ContainerStarted","Data":"1ed1a2167332877b03c55ee5b2215fb47721327084ac99b337b241158b70202b"} Oct 03 16:42:45 crc kubenswrapper[4744]: I1003 16:42:45.897092 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"30c077c9e7442ebfd1ca7f737dd9e680328a67753a20e26795bd1808da8a069b"} Oct 03 16:42:45 crc kubenswrapper[4744]: I1003 16:42:45.897133 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"c72fdbc1ee01b5c4a4f7afe5e8b010d3e4cf22f524e6972b4644cfe18efd46b6"} Oct 03 16:42:45 crc kubenswrapper[4744]: I1003 16:42:45.897151 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"aca9fe14b14fba867e7125bc2cbe65969482e64fa0815796615e87860f64172f"} Oct 03 16:42:46 crc kubenswrapper[4744]: I1003 16:42:46.910121 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="139159c5-0601-42e2-b71f-1497543df40f" path="/var/lib/kubelet/pods/139159c5-0601-42e2-b71f-1497543df40f/volumes" Oct 03 16:42:46 crc kubenswrapper[4744]: I1003 16:42:46.913791 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"fc54443f205420f2296213a3e1900fa91b135211a02568491e6d9d07389f4ea3"} Oct 03 16:42:47 crc kubenswrapper[4744]: I1003 16:42:47.252674 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5696-account-create-jn6hh" Oct 03 16:42:47 crc kubenswrapper[4744]: I1003 16:42:47.299539 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsj2w\" (UniqueName: \"kubernetes.io/projected/d34f7f18-2333-4975-9606-9dd4ec282f3e-kube-api-access-nsj2w\") pod \"d34f7f18-2333-4975-9606-9dd4ec282f3e\" (UID: \"d34f7f18-2333-4975-9606-9dd4ec282f3e\") " Oct 03 16:42:47 crc kubenswrapper[4744]: I1003 16:42:47.308375 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d34f7f18-2333-4975-9606-9dd4ec282f3e-kube-api-access-nsj2w" (OuterVolumeSpecName: "kube-api-access-nsj2w") pod "d34f7f18-2333-4975-9606-9dd4ec282f3e" (UID: "d34f7f18-2333-4975-9606-9dd4ec282f3e"). InnerVolumeSpecName "kube-api-access-nsj2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:42:47 crc kubenswrapper[4744]: I1003 16:42:47.401591 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsj2w\" (UniqueName: \"kubernetes.io/projected/d34f7f18-2333-4975-9606-9dd4ec282f3e-kube-api-access-nsj2w\") on node \"crc\" DevicePath \"\"" Oct 03 16:42:47 crc kubenswrapper[4744]: I1003 16:42:47.932396 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5696-account-create-jn6hh" event={"ID":"d34f7f18-2333-4975-9606-9dd4ec282f3e","Type":"ContainerDied","Data":"1ed1a2167332877b03c55ee5b2215fb47721327084ac99b337b241158b70202b"} Oct 03 16:42:47 crc kubenswrapper[4744]: I1003 16:42:47.932705 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ed1a2167332877b03c55ee5b2215fb47721327084ac99b337b241158b70202b" Oct 03 16:42:47 crc kubenswrapper[4744]: I1003 16:42:47.932470 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5696-account-create-jn6hh" Oct 03 16:42:48 crc kubenswrapper[4744]: I1003 16:42:48.945527 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"e0f1126cfb2c53b2efa0fa3b916092a46f4c2fa083b73188cf625b0355c6055f"} Oct 03 16:42:48 crc kubenswrapper[4744]: I1003 16:42:48.945568 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"7ca9790a71bc89250135c7f0f56022d15f9ba64c8496d90690fb972297da455b"} Oct 03 16:42:48 crc kubenswrapper[4744]: I1003 16:42:48.945578 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"f1611cb484ee2c23d76499eed6748da4d4dbb6b2c583828d205312280d3e4ea8"} Oct 03 16:42:48 crc kubenswrapper[4744]: I1003 16:42:48.945585 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"77d8dbf5c36ded6ca27329fbe7f560f93f61658ed1aa0c0a74c266423c4d7cff"} Oct 03 16:42:48 crc kubenswrapper[4744]: I1003 16:42:48.945593 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"693b215ea688466d4b1f173cde8afe103bf6975a5858e5f7a608c259d9fc4ebc"} Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.378837 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-l78wg"] Oct 03 16:42:49 crc kubenswrapper[4744]: E1003 16:42:49.379194 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="139159c5-0601-42e2-b71f-1497543df40f" containerName="ovn-config" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.379206 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="139159c5-0601-42e2-b71f-1497543df40f" containerName="ovn-config" Oct 03 16:42:49 crc kubenswrapper[4744]: E1003 16:42:49.379225 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d34f7f18-2333-4975-9606-9dd4ec282f3e" containerName="mariadb-account-create" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.379230 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d34f7f18-2333-4975-9606-9dd4ec282f3e" containerName="mariadb-account-create" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.379379 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="139159c5-0601-42e2-b71f-1497543df40f" containerName="ovn-config" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.379399 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d34f7f18-2333-4975-9606-9dd4ec282f3e" containerName="mariadb-account-create" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.379969 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l78wg" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.385174 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.385380 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-fk4vm" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.394886 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-l78wg"] Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.435760 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-db-sync-config-data\") pod \"glance-db-sync-l78wg\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " pod="openstack/glance-db-sync-l78wg" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.435805 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-config-data\") pod \"glance-db-sync-l78wg\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " pod="openstack/glance-db-sync-l78wg" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.435849 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7kfz\" (UniqueName: \"kubernetes.io/projected/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-kube-api-access-n7kfz\") pod \"glance-db-sync-l78wg\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " pod="openstack/glance-db-sync-l78wg" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.435888 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-combined-ca-bundle\") pod \"glance-db-sync-l78wg\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " pod="openstack/glance-db-sync-l78wg" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.537829 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-combined-ca-bundle\") pod \"glance-db-sync-l78wg\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " pod="openstack/glance-db-sync-l78wg" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.537948 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-db-sync-config-data\") pod \"glance-db-sync-l78wg\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " pod="openstack/glance-db-sync-l78wg" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.537967 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-config-data\") pod \"glance-db-sync-l78wg\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " pod="openstack/glance-db-sync-l78wg" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.538011 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7kfz\" (UniqueName: \"kubernetes.io/projected/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-kube-api-access-n7kfz\") pod \"glance-db-sync-l78wg\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " pod="openstack/glance-db-sync-l78wg" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.544245 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-config-data\") pod \"glance-db-sync-l78wg\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " pod="openstack/glance-db-sync-l78wg" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.544245 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-combined-ca-bundle\") pod \"glance-db-sync-l78wg\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " pod="openstack/glance-db-sync-l78wg" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.546726 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-db-sync-config-data\") pod \"glance-db-sync-l78wg\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " pod="openstack/glance-db-sync-l78wg" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.565956 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7kfz\" (UniqueName: \"kubernetes.io/projected/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-kube-api-access-n7kfz\") pod \"glance-db-sync-l78wg\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " pod="openstack/glance-db-sync-l78wg" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.707546 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l78wg" Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.962446 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"3313bc884b48708f7a757b43463069e26f06e529f82c2341a84f93e592098eb0"} Oct 03 16:42:49 crc kubenswrapper[4744]: I1003 16:42:49.962487 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"133725f5-8ac3-4f45-b6c3-f7aea913711d","Type":"ContainerStarted","Data":"6ef3ca2fdd4ca06e5017114a292580f9e8aeaefb4bc65c50a8f8e3a68a0b78f0"} Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.003129 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=19.98273514 podStartE2EDuration="29.003110607s" podCreationTimestamp="2025-10-03 16:42:21 +0000 UTC" firstStartedPulling="2025-10-03 16:42:38.898697368 +0000 UTC m=+1085.178573274" lastFinishedPulling="2025-10-03 16:42:47.919072845 +0000 UTC m=+1094.198948741" observedRunningTime="2025-10-03 16:42:49.999325769 +0000 UTC m=+1096.279201675" watchObservedRunningTime="2025-10-03 16:42:50.003110607 +0000 UTC m=+1096.282986503" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.261958 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-l78wg"] Oct 03 16:42:50 crc kubenswrapper[4744]: W1003 16:42:50.267885 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d4a99fa_d668_4798_b988_a8d8fb9e3b28.slice/crio-11b4e4ac9e3f5bcfedc0ab3203c3a000d16467eadb4aaf1f4845d3f438543c69 WatchSource:0}: Error finding container 11b4e4ac9e3f5bcfedc0ab3203c3a000d16467eadb4aaf1f4845d3f438543c69: Status 404 returned error can't find the container with id 11b4e4ac9e3f5bcfedc0ab3203c3a000d16467eadb4aaf1f4845d3f438543c69 Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.292843 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-8hrkj"] Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.294592 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.296634 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.318620 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-8hrkj"] Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.354064 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-config\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.354163 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.354243 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.354311 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q52cx\" (UniqueName: \"kubernetes.io/projected/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-kube-api-access-q52cx\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.354410 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.354463 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.455604 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.455649 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.455671 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q52cx\" (UniqueName: \"kubernetes.io/projected/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-kube-api-access-q52cx\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.455705 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.455724 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.455820 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-config\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.456684 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-config\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.456754 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.456775 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.457029 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.460041 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.484656 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q52cx\" (UniqueName: \"kubernetes.io/projected/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-kube-api-access-q52cx\") pod \"dnsmasq-dns-77585f5f8c-8hrkj\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.645220 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:50 crc kubenswrapper[4744]: I1003 16:42:50.970149 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l78wg" event={"ID":"3d4a99fa-d668-4798-b988-a8d8fb9e3b28","Type":"ContainerStarted","Data":"11b4e4ac9e3f5bcfedc0ab3203c3a000d16467eadb4aaf1f4845d3f438543c69"} Oct 03 16:42:51 crc kubenswrapper[4744]: I1003 16:42:51.110042 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-8hrkj"] Oct 03 16:42:51 crc kubenswrapper[4744]: W1003 16:42:51.113685 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c4651bb_32f8_4baa_b0a3_a8fd37ca9914.slice/crio-505cedd02078765974fb0cf966b86f4451231120d88a429acc9c92cbd2a8dd38 WatchSource:0}: Error finding container 505cedd02078765974fb0cf966b86f4451231120d88a429acc9c92cbd2a8dd38: Status 404 returned error can't find the container with id 505cedd02078765974fb0cf966b86f4451231120d88a429acc9c92cbd2a8dd38 Oct 03 16:42:51 crc kubenswrapper[4744]: I1003 16:42:51.981705 4744 generic.go:334] "Generic (PLEG): container finished" podID="3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" containerID="e953d762fe5d2bd3ded9de739a090ecbb8970d7fe582e876cf8a5fc71aea20f1" exitCode=0 Oct 03 16:42:51 crc kubenswrapper[4744]: I1003 16:42:51.982201 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" event={"ID":"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914","Type":"ContainerDied","Data":"e953d762fe5d2bd3ded9de739a090ecbb8970d7fe582e876cf8a5fc71aea20f1"} Oct 03 16:42:51 crc kubenswrapper[4744]: I1003 16:42:51.982230 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" event={"ID":"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914","Type":"ContainerStarted","Data":"505cedd02078765974fb0cf966b86f4451231120d88a429acc9c92cbd2a8dd38"} Oct 03 16:42:52 crc kubenswrapper[4744]: I1003 16:42:52.993205 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" event={"ID":"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914","Type":"ContainerStarted","Data":"1228505882f8ef70a3feeb4a8aecc382e7742259e6fb684543f703f4163038a5"} Oct 03 16:42:52 crc kubenswrapper[4744]: I1003 16:42:52.993756 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:42:53 crc kubenswrapper[4744]: I1003 16:42:53.025122 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" podStartSLOduration=3.025104692 podStartE2EDuration="3.025104692s" podCreationTimestamp="2025-10-03 16:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:42:53.019314292 +0000 UTC m=+1099.299190188" watchObservedRunningTime="2025-10-03 16:42:53.025104692 +0000 UTC m=+1099.304980578" Oct 03 16:42:55 crc kubenswrapper[4744]: I1003 16:42:55.652718 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 03 16:42:55 crc kubenswrapper[4744]: I1003 16:42:55.924787 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.082373 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-6j7qv"] Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.083732 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6j7qv" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.099943 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6j7qv"] Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.158282 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8jlw\" (UniqueName: \"kubernetes.io/projected/019c5b16-937c-45d2-9003-503f4d88995d-kube-api-access-r8jlw\") pod \"cinder-db-create-6j7qv\" (UID: \"019c5b16-937c-45d2-9003-503f4d88995d\") " pod="openstack/cinder-db-create-6j7qv" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.180507 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-p5bkm"] Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.181764 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-p5bkm" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.188565 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-p5bkm"] Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.243620 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-cpdb5"] Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.244583 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cpdb5" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.252433 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.253164 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.253523 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2z2zg" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.254154 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.257057 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-cpdb5"] Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.259393 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-config-data\") pod \"keystone-db-sync-cpdb5\" (UID: \"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe\") " pod="openstack/keystone-db-sync-cpdb5" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.259447 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64lcw\" (UniqueName: \"kubernetes.io/projected/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-kube-api-access-64lcw\") pod \"keystone-db-sync-cpdb5\" (UID: \"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe\") " pod="openstack/keystone-db-sync-cpdb5" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.259486 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8jlw\" (UniqueName: \"kubernetes.io/projected/019c5b16-937c-45d2-9003-503f4d88995d-kube-api-access-r8jlw\") pod \"cinder-db-create-6j7qv\" (UID: \"019c5b16-937c-45d2-9003-503f4d88995d\") " pod="openstack/cinder-db-create-6j7qv" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.259529 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-combined-ca-bundle\") pod \"keystone-db-sync-cpdb5\" (UID: \"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe\") " pod="openstack/keystone-db-sync-cpdb5" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.259559 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vzvd\" (UniqueName: \"kubernetes.io/projected/97fcb340-8304-4bae-9d2b-f0b04dcea8bf-kube-api-access-5vzvd\") pod \"barbican-db-create-p5bkm\" (UID: \"97fcb340-8304-4bae-9d2b-f0b04dcea8bf\") " pod="openstack/barbican-db-create-p5bkm" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.292784 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-wdg7r"] Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.295425 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8jlw\" (UniqueName: \"kubernetes.io/projected/019c5b16-937c-45d2-9003-503f4d88995d-kube-api-access-r8jlw\") pod \"cinder-db-create-6j7qv\" (UID: \"019c5b16-937c-45d2-9003-503f4d88995d\") " pod="openstack/cinder-db-create-6j7qv" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.295786 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-wdg7r" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.301026 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-wdg7r"] Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.383564 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-config-data\") pod \"keystone-db-sync-cpdb5\" (UID: \"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe\") " pod="openstack/keystone-db-sync-cpdb5" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.383682 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64lcw\" (UniqueName: \"kubernetes.io/projected/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-kube-api-access-64lcw\") pod \"keystone-db-sync-cpdb5\" (UID: \"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe\") " pod="openstack/keystone-db-sync-cpdb5" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.383805 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-combined-ca-bundle\") pod \"keystone-db-sync-cpdb5\" (UID: \"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe\") " pod="openstack/keystone-db-sync-cpdb5" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.383875 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rrrt\" (UniqueName: \"kubernetes.io/projected/18c69bbd-4199-4da7-8e11-e05ea8680342-kube-api-access-8rrrt\") pod \"manila-db-create-wdg7r\" (UID: \"18c69bbd-4199-4da7-8e11-e05ea8680342\") " pod="openstack/manila-db-create-wdg7r" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.383909 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vzvd\" (UniqueName: \"kubernetes.io/projected/97fcb340-8304-4bae-9d2b-f0b04dcea8bf-kube-api-access-5vzvd\") pod \"barbican-db-create-p5bkm\" (UID: \"97fcb340-8304-4bae-9d2b-f0b04dcea8bf\") " pod="openstack/barbican-db-create-p5bkm" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.393105 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-combined-ca-bundle\") pod \"keystone-db-sync-cpdb5\" (UID: \"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe\") " pod="openstack/keystone-db-sync-cpdb5" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.401339 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-config-data\") pod \"keystone-db-sync-cpdb5\" (UID: \"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe\") " pod="openstack/keystone-db-sync-cpdb5" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.401747 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64lcw\" (UniqueName: \"kubernetes.io/projected/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-kube-api-access-64lcw\") pod \"keystone-db-sync-cpdb5\" (UID: \"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe\") " pod="openstack/keystone-db-sync-cpdb5" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.410837 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6j7qv" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.411136 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vzvd\" (UniqueName: \"kubernetes.io/projected/97fcb340-8304-4bae-9d2b-f0b04dcea8bf-kube-api-access-5vzvd\") pod \"barbican-db-create-p5bkm\" (UID: \"97fcb340-8304-4bae-9d2b-f0b04dcea8bf\") " pod="openstack/barbican-db-create-p5bkm" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.481124 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-xrtkk"] Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.486065 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rrrt\" (UniqueName: \"kubernetes.io/projected/18c69bbd-4199-4da7-8e11-e05ea8680342-kube-api-access-8rrrt\") pod \"manila-db-create-wdg7r\" (UID: \"18c69bbd-4199-4da7-8e11-e05ea8680342\") " pod="openstack/manila-db-create-wdg7r" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.487168 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xrtkk" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.499063 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xrtkk"] Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.510431 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-p5bkm" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.520756 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rrrt\" (UniqueName: \"kubernetes.io/projected/18c69bbd-4199-4da7-8e11-e05ea8680342-kube-api-access-8rrrt\") pod \"manila-db-create-wdg7r\" (UID: \"18c69bbd-4199-4da7-8e11-e05ea8680342\") " pod="openstack/manila-db-create-wdg7r" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.565202 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cpdb5" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.589562 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcs8d\" (UniqueName: \"kubernetes.io/projected/299d80eb-fc43-4647-83be-a792412df7b1-kube-api-access-vcs8d\") pod \"neutron-db-create-xrtkk\" (UID: \"299d80eb-fc43-4647-83be-a792412df7b1\") " pod="openstack/neutron-db-create-xrtkk" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.645821 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-wdg7r" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.691232 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcs8d\" (UniqueName: \"kubernetes.io/projected/299d80eb-fc43-4647-83be-a792412df7b1-kube-api-access-vcs8d\") pod \"neutron-db-create-xrtkk\" (UID: \"299d80eb-fc43-4647-83be-a792412df7b1\") " pod="openstack/neutron-db-create-xrtkk" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.710105 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcs8d\" (UniqueName: \"kubernetes.io/projected/299d80eb-fc43-4647-83be-a792412df7b1-kube-api-access-vcs8d\") pod \"neutron-db-create-xrtkk\" (UID: \"299d80eb-fc43-4647-83be-a792412df7b1\") " pod="openstack/neutron-db-create-xrtkk" Oct 03 16:42:56 crc kubenswrapper[4744]: I1003 16:42:56.808600 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xrtkk" Oct 03 16:43:00 crc kubenswrapper[4744]: I1003 16:43:00.646649 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:43:00 crc kubenswrapper[4744]: I1003 16:43:00.710712 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-g6wjt"] Oct 03 16:43:00 crc kubenswrapper[4744]: I1003 16:43:00.712067 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-g6wjt" podUID="a01c5ef1-aeea-431e-a744-2be2b1a28acf" containerName="dnsmasq-dns" containerID="cri-o://8bc1ed4255353772b0f525b2c50735fff42a10b152105eb81c01847286ad05af" gracePeriod=10 Oct 03 16:43:01 crc kubenswrapper[4744]: I1003 16:43:01.073439 4744 generic.go:334] "Generic (PLEG): container finished" podID="a01c5ef1-aeea-431e-a744-2be2b1a28acf" containerID="8bc1ed4255353772b0f525b2c50735fff42a10b152105eb81c01847286ad05af" exitCode=0 Oct 03 16:43:01 crc kubenswrapper[4744]: I1003 16:43:01.073529 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-g6wjt" event={"ID":"a01c5ef1-aeea-431e-a744-2be2b1a28acf","Type":"ContainerDied","Data":"8bc1ed4255353772b0f525b2c50735fff42a10b152105eb81c01847286ad05af"} Oct 03 16:43:01 crc kubenswrapper[4744]: I1003 16:43:01.385524 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-g6wjt" podUID="a01c5ef1-aeea-431e-a744-2be2b1a28acf" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.091446 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-g6wjt" event={"ID":"a01c5ef1-aeea-431e-a744-2be2b1a28acf","Type":"ContainerDied","Data":"80ba143779351234eb3d494bbd7b2e43a39d3ee313d01f2fc545e8a8f86ccef0"} Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.092060 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80ba143779351234eb3d494bbd7b2e43a39d3ee313d01f2fc545e8a8f86ccef0" Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.109543 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.206996 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6xkl\" (UniqueName: \"kubernetes.io/projected/a01c5ef1-aeea-431e-a744-2be2b1a28acf-kube-api-access-d6xkl\") pod \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.207095 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-ovsdbserver-nb\") pod \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.207193 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-ovsdbserver-sb\") pod \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.207292 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-dns-svc\") pod \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.207353 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-config\") pod \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\" (UID: \"a01c5ef1-aeea-431e-a744-2be2b1a28acf\") " Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.245625 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a01c5ef1-aeea-431e-a744-2be2b1a28acf-kube-api-access-d6xkl" (OuterVolumeSpecName: "kube-api-access-d6xkl") pod "a01c5ef1-aeea-431e-a744-2be2b1a28acf" (UID: "a01c5ef1-aeea-431e-a744-2be2b1a28acf"). InnerVolumeSpecName "kube-api-access-d6xkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.310950 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6xkl\" (UniqueName: \"kubernetes.io/projected/a01c5ef1-aeea-431e-a744-2be2b1a28acf-kube-api-access-d6xkl\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.336228 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a01c5ef1-aeea-431e-a744-2be2b1a28acf" (UID: "a01c5ef1-aeea-431e-a744-2be2b1a28acf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.348118 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a01c5ef1-aeea-431e-a744-2be2b1a28acf" (UID: "a01c5ef1-aeea-431e-a744-2be2b1a28acf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.357973 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a01c5ef1-aeea-431e-a744-2be2b1a28acf" (UID: "a01c5ef1-aeea-431e-a744-2be2b1a28acf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.373008 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-config" (OuterVolumeSpecName: "config") pod "a01c5ef1-aeea-431e-a744-2be2b1a28acf" (UID: "a01c5ef1-aeea-431e-a744-2be2b1a28acf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:03 crc kubenswrapper[4744]: W1003 16:43:03.382228 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod019c5b16_937c_45d2_9003_503f4d88995d.slice/crio-7b26847541092d9fe6c0e3cb2591a7fef3e0150d922047cfbcb82027941807e6 WatchSource:0}: Error finding container 7b26847541092d9fe6c0e3cb2591a7fef3e0150d922047cfbcb82027941807e6: Status 404 returned error can't find the container with id 7b26847541092d9fe6c0e3cb2591a7fef3e0150d922047cfbcb82027941807e6 Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.382302 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6j7qv"] Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.412673 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.413087 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.413098 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.413108 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a01c5ef1-aeea-431e-a744-2be2b1a28acf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.622536 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xrtkk"] Oct 03 16:43:03 crc kubenswrapper[4744]: W1003 16:43:03.632554 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod299d80eb_fc43_4647_83be_a792412df7b1.slice/crio-d24afb36e4338ceb09f394daadfd68a2f6f8b64c37cae0046c3de440ecb4ee34 WatchSource:0}: Error finding container d24afb36e4338ceb09f394daadfd68a2f6f8b64c37cae0046c3de440ecb4ee34: Status 404 returned error can't find the container with id d24afb36e4338ceb09f394daadfd68a2f6f8b64c37cae0046c3de440ecb4ee34 Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.632755 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-p5bkm"] Oct 03 16:43:03 crc kubenswrapper[4744]: W1003 16:43:03.634946 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97fcb340_8304_4bae_9d2b_f0b04dcea8bf.slice/crio-39b6891b8504cc652ee2307668c7e64aebf3f6130f7ab1f83281a55fa7abc0c4 WatchSource:0}: Error finding container 39b6891b8504cc652ee2307668c7e64aebf3f6130f7ab1f83281a55fa7abc0c4: Status 404 returned error can't find the container with id 39b6891b8504cc652ee2307668c7e64aebf3f6130f7ab1f83281a55fa7abc0c4 Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.732326 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-wdg7r"] Oct 03 16:43:03 crc kubenswrapper[4744]: W1003 16:43:03.735287 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd5b11ee_b171_4bd8_9f40_34d9dcb428fe.slice/crio-93e44c57540d1613f4a2b39ed0895239e15a08ab2d9beadf4526398a8ff48418 WatchSource:0}: Error finding container 93e44c57540d1613f4a2b39ed0895239e15a08ab2d9beadf4526398a8ff48418: Status 404 returned error can't find the container with id 93e44c57540d1613f4a2b39ed0895239e15a08ab2d9beadf4526398a8ff48418 Oct 03 16:43:03 crc kubenswrapper[4744]: I1003 16:43:03.737850 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-cpdb5"] Oct 03 16:43:03 crc kubenswrapper[4744]: W1003 16:43:03.742096 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18c69bbd_4199_4da7_8e11_e05ea8680342.slice/crio-c26d1825942aafd6e070559eba1cc5dce81bce5226066cedac81217d6b9c523b WatchSource:0}: Error finding container c26d1825942aafd6e070559eba1cc5dce81bce5226066cedac81217d6b9c523b: Status 404 returned error can't find the container with id c26d1825942aafd6e070559eba1cc5dce81bce5226066cedac81217d6b9c523b Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.105207 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-wdg7r" event={"ID":"18c69bbd-4199-4da7-8e11-e05ea8680342","Type":"ContainerStarted","Data":"04dc55f54d87503af16952d67d9653f221b6abe42f04b82718de2e1d235511f5"} Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.105572 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-wdg7r" event={"ID":"18c69bbd-4199-4da7-8e11-e05ea8680342","Type":"ContainerStarted","Data":"c26d1825942aafd6e070559eba1cc5dce81bce5226066cedac81217d6b9c523b"} Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.110228 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l78wg" event={"ID":"3d4a99fa-d668-4798-b988-a8d8fb9e3b28","Type":"ContainerStarted","Data":"0991e3d7331305f7c44f8d526bf5753a3739f1d50ac22a91c69869006166c4c1"} Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.112648 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cpdb5" event={"ID":"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe","Type":"ContainerStarted","Data":"93e44c57540d1613f4a2b39ed0895239e15a08ab2d9beadf4526398a8ff48418"} Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.114734 4744 generic.go:334] "Generic (PLEG): container finished" podID="019c5b16-937c-45d2-9003-503f4d88995d" containerID="40905c154d75a80574ac970ee03bdfa5e8f403139e5cb5aef9cc275934ef0411" exitCode=0 Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.114781 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6j7qv" event={"ID":"019c5b16-937c-45d2-9003-503f4d88995d","Type":"ContainerDied","Data":"40905c154d75a80574ac970ee03bdfa5e8f403139e5cb5aef9cc275934ef0411"} Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.114844 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6j7qv" event={"ID":"019c5b16-937c-45d2-9003-503f4d88995d","Type":"ContainerStarted","Data":"7b26847541092d9fe6c0e3cb2591a7fef3e0150d922047cfbcb82027941807e6"} Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.116154 4744 generic.go:334] "Generic (PLEG): container finished" podID="97fcb340-8304-4bae-9d2b-f0b04dcea8bf" containerID="798f39b84e71fc24d78389f8bd3f1c41309abc68525621cb19513b377551f6e9" exitCode=0 Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.116300 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-p5bkm" event={"ID":"97fcb340-8304-4bae-9d2b-f0b04dcea8bf","Type":"ContainerDied","Data":"798f39b84e71fc24d78389f8bd3f1c41309abc68525621cb19513b377551f6e9"} Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.116380 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-p5bkm" event={"ID":"97fcb340-8304-4bae-9d2b-f0b04dcea8bf","Type":"ContainerStarted","Data":"39b6891b8504cc652ee2307668c7e64aebf3f6130f7ab1f83281a55fa7abc0c4"} Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.117459 4744 generic.go:334] "Generic (PLEG): container finished" podID="299d80eb-fc43-4647-83be-a792412df7b1" containerID="6ed89bc00a03e9a367db4ca1449835b8957b051b02546da6ccde505e97e3630c" exitCode=0 Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.117619 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-g6wjt" Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.117684 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xrtkk" event={"ID":"299d80eb-fc43-4647-83be-a792412df7b1","Type":"ContainerDied","Data":"6ed89bc00a03e9a367db4ca1449835b8957b051b02546da6ccde505e97e3630c"} Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.117716 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xrtkk" event={"ID":"299d80eb-fc43-4647-83be-a792412df7b1","Type":"ContainerStarted","Data":"d24afb36e4338ceb09f394daadfd68a2f6f8b64c37cae0046c3de440ecb4ee34"} Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.151975 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-create-wdg7r" podStartSLOduration=8.151947022 podStartE2EDuration="8.151947022s" podCreationTimestamp="2025-10-03 16:42:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:04.127330604 +0000 UTC m=+1110.407206500" watchObservedRunningTime="2025-10-03 16:43:04.151947022 +0000 UTC m=+1110.431822958" Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.195668 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-l78wg" podStartSLOduration=2.482293233 podStartE2EDuration="15.195649073s" podCreationTimestamp="2025-10-03 16:42:49 +0000 UTC" firstStartedPulling="2025-10-03 16:42:50.271247139 +0000 UTC m=+1096.551123035" lastFinishedPulling="2025-10-03 16:43:02.984602979 +0000 UTC m=+1109.264478875" observedRunningTime="2025-10-03 16:43:04.180726114 +0000 UTC m=+1110.460602030" watchObservedRunningTime="2025-10-03 16:43:04.195649073 +0000 UTC m=+1110.475524979" Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.221478 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-g6wjt"] Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.226938 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-g6wjt"] Oct 03 16:43:04 crc kubenswrapper[4744]: I1003 16:43:04.905233 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a01c5ef1-aeea-431e-a744-2be2b1a28acf" path="/var/lib/kubelet/pods/a01c5ef1-aeea-431e-a744-2be2b1a28acf/volumes" Oct 03 16:43:05 crc kubenswrapper[4744]: I1003 16:43:05.128885 4744 generic.go:334] "Generic (PLEG): container finished" podID="18c69bbd-4199-4da7-8e11-e05ea8680342" containerID="04dc55f54d87503af16952d67d9653f221b6abe42f04b82718de2e1d235511f5" exitCode=0 Oct 03 16:43:05 crc kubenswrapper[4744]: I1003 16:43:05.128938 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-wdg7r" event={"ID":"18c69bbd-4199-4da7-8e11-e05ea8680342","Type":"ContainerDied","Data":"04dc55f54d87503af16952d67d9653f221b6abe42f04b82718de2e1d235511f5"} Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.019690 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-p5bkm" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.026312 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6j7qv" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.037924 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xrtkk" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.073023 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-wdg7r" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.095138 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8jlw\" (UniqueName: \"kubernetes.io/projected/019c5b16-937c-45d2-9003-503f4d88995d-kube-api-access-r8jlw\") pod \"019c5b16-937c-45d2-9003-503f4d88995d\" (UID: \"019c5b16-937c-45d2-9003-503f4d88995d\") " Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.095230 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vzvd\" (UniqueName: \"kubernetes.io/projected/97fcb340-8304-4bae-9d2b-f0b04dcea8bf-kube-api-access-5vzvd\") pod \"97fcb340-8304-4bae-9d2b-f0b04dcea8bf\" (UID: \"97fcb340-8304-4bae-9d2b-f0b04dcea8bf\") " Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.095337 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rrrt\" (UniqueName: \"kubernetes.io/projected/18c69bbd-4199-4da7-8e11-e05ea8680342-kube-api-access-8rrrt\") pod \"18c69bbd-4199-4da7-8e11-e05ea8680342\" (UID: \"18c69bbd-4199-4da7-8e11-e05ea8680342\") " Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.095386 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcs8d\" (UniqueName: \"kubernetes.io/projected/299d80eb-fc43-4647-83be-a792412df7b1-kube-api-access-vcs8d\") pod \"299d80eb-fc43-4647-83be-a792412df7b1\" (UID: \"299d80eb-fc43-4647-83be-a792412df7b1\") " Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.109306 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/299d80eb-fc43-4647-83be-a792412df7b1-kube-api-access-vcs8d" (OuterVolumeSpecName: "kube-api-access-vcs8d") pod "299d80eb-fc43-4647-83be-a792412df7b1" (UID: "299d80eb-fc43-4647-83be-a792412df7b1"). InnerVolumeSpecName "kube-api-access-vcs8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.111997 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97fcb340-8304-4bae-9d2b-f0b04dcea8bf-kube-api-access-5vzvd" (OuterVolumeSpecName: "kube-api-access-5vzvd") pod "97fcb340-8304-4bae-9d2b-f0b04dcea8bf" (UID: "97fcb340-8304-4bae-9d2b-f0b04dcea8bf"). InnerVolumeSpecName "kube-api-access-5vzvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.112482 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18c69bbd-4199-4da7-8e11-e05ea8680342-kube-api-access-8rrrt" (OuterVolumeSpecName: "kube-api-access-8rrrt") pod "18c69bbd-4199-4da7-8e11-e05ea8680342" (UID: "18c69bbd-4199-4da7-8e11-e05ea8680342"). InnerVolumeSpecName "kube-api-access-8rrrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.114723 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/019c5b16-937c-45d2-9003-503f4d88995d-kube-api-access-r8jlw" (OuterVolumeSpecName: "kube-api-access-r8jlw") pod "019c5b16-937c-45d2-9003-503f4d88995d" (UID: "019c5b16-937c-45d2-9003-503f4d88995d"). InnerVolumeSpecName "kube-api-access-r8jlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.155898 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6j7qv" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.156134 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6j7qv" event={"ID":"019c5b16-937c-45d2-9003-503f4d88995d","Type":"ContainerDied","Data":"7b26847541092d9fe6c0e3cb2591a7fef3e0150d922047cfbcb82027941807e6"} Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.156177 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b26847541092d9fe6c0e3cb2591a7fef3e0150d922047cfbcb82027941807e6" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.159458 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-p5bkm" event={"ID":"97fcb340-8304-4bae-9d2b-f0b04dcea8bf","Type":"ContainerDied","Data":"39b6891b8504cc652ee2307668c7e64aebf3f6130f7ab1f83281a55fa7abc0c4"} Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.159509 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39b6891b8504cc652ee2307668c7e64aebf3f6130f7ab1f83281a55fa7abc0c4" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.159562 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-p5bkm" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.161231 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xrtkk" event={"ID":"299d80eb-fc43-4647-83be-a792412df7b1","Type":"ContainerDied","Data":"d24afb36e4338ceb09f394daadfd68a2f6f8b64c37cae0046c3de440ecb4ee34"} Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.161303 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d24afb36e4338ceb09f394daadfd68a2f6f8b64c37cae0046c3de440ecb4ee34" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.161365 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xrtkk" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.167278 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-wdg7r" event={"ID":"18c69bbd-4199-4da7-8e11-e05ea8680342","Type":"ContainerDied","Data":"c26d1825942aafd6e070559eba1cc5dce81bce5226066cedac81217d6b9c523b"} Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.167322 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c26d1825942aafd6e070559eba1cc5dce81bce5226066cedac81217d6b9c523b" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.167387 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-wdg7r" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.197617 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rrrt\" (UniqueName: \"kubernetes.io/projected/18c69bbd-4199-4da7-8e11-e05ea8680342-kube-api-access-8rrrt\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.197646 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcs8d\" (UniqueName: \"kubernetes.io/projected/299d80eb-fc43-4647-83be-a792412df7b1-kube-api-access-vcs8d\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.197657 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8jlw\" (UniqueName: \"kubernetes.io/projected/019c5b16-937c-45d2-9003-503f4d88995d-kube-api-access-r8jlw\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:08 crc kubenswrapper[4744]: I1003 16:43:08.197668 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vzvd\" (UniqueName: \"kubernetes.io/projected/97fcb340-8304-4bae-9d2b-f0b04dcea8bf-kube-api-access-5vzvd\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:09 crc kubenswrapper[4744]: I1003 16:43:09.175605 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cpdb5" event={"ID":"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe","Type":"ContainerStarted","Data":"c9f5c9f33347190f940722acea6862eb028c2c50d98422eac5885f8ea9f40616"} Oct 03 16:43:09 crc kubenswrapper[4744]: I1003 16:43:09.198384 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-cpdb5" podStartSLOduration=9.116124858 podStartE2EDuration="13.198365948s" podCreationTimestamp="2025-10-03 16:42:56 +0000 UTC" firstStartedPulling="2025-10-03 16:43:03.737700341 +0000 UTC m=+1110.017576237" lastFinishedPulling="2025-10-03 16:43:07.819941411 +0000 UTC m=+1114.099817327" observedRunningTime="2025-10-03 16:43:09.190663983 +0000 UTC m=+1115.470539879" watchObservedRunningTime="2025-10-03 16:43:09.198365948 +0000 UTC m=+1115.478241844" Oct 03 16:43:10 crc kubenswrapper[4744]: I1003 16:43:10.190227 4744 generic.go:334] "Generic (PLEG): container finished" podID="3d4a99fa-d668-4798-b988-a8d8fb9e3b28" containerID="0991e3d7331305f7c44f8d526bf5753a3739f1d50ac22a91c69869006166c4c1" exitCode=0 Oct 03 16:43:10 crc kubenswrapper[4744]: I1003 16:43:10.190287 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l78wg" event={"ID":"3d4a99fa-d668-4798-b988-a8d8fb9e3b28","Type":"ContainerDied","Data":"0991e3d7331305f7c44f8d526bf5753a3739f1d50ac22a91c69869006166c4c1"} Oct 03 16:43:10 crc kubenswrapper[4744]: E1003 16:43:10.990233 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd5b11ee_b171_4bd8_9f40_34d9dcb428fe.slice/crio-conmon-c9f5c9f33347190f940722acea6862eb028c2c50d98422eac5885f8ea9f40616.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd5b11ee_b171_4bd8_9f40_34d9dcb428fe.slice/crio-c9f5c9f33347190f940722acea6862eb028c2c50d98422eac5885f8ea9f40616.scope\": RecentStats: unable to find data in memory cache]" Oct 03 16:43:11 crc kubenswrapper[4744]: I1003 16:43:11.203489 4744 generic.go:334] "Generic (PLEG): container finished" podID="cd5b11ee-b171-4bd8-9f40-34d9dcb428fe" containerID="c9f5c9f33347190f940722acea6862eb028c2c50d98422eac5885f8ea9f40616" exitCode=0 Oct 03 16:43:11 crc kubenswrapper[4744]: I1003 16:43:11.203752 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cpdb5" event={"ID":"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe","Type":"ContainerDied","Data":"c9f5c9f33347190f940722acea6862eb028c2c50d98422eac5885f8ea9f40616"} Oct 03 16:43:11 crc kubenswrapper[4744]: I1003 16:43:11.663707 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l78wg" Oct 03 16:43:11 crc kubenswrapper[4744]: I1003 16:43:11.787068 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-config-data\") pod \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " Oct 03 16:43:11 crc kubenswrapper[4744]: I1003 16:43:11.787130 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-db-sync-config-data\") pod \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " Oct 03 16:43:11 crc kubenswrapper[4744]: I1003 16:43:11.787153 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7kfz\" (UniqueName: \"kubernetes.io/projected/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-kube-api-access-n7kfz\") pod \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " Oct 03 16:43:11 crc kubenswrapper[4744]: I1003 16:43:11.787252 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-combined-ca-bundle\") pod \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\" (UID: \"3d4a99fa-d668-4798-b988-a8d8fb9e3b28\") " Oct 03 16:43:11 crc kubenswrapper[4744]: I1003 16:43:11.792285 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3d4a99fa-d668-4798-b988-a8d8fb9e3b28" (UID: "3d4a99fa-d668-4798-b988-a8d8fb9e3b28"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:11 crc kubenswrapper[4744]: I1003 16:43:11.792359 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-kube-api-access-n7kfz" (OuterVolumeSpecName: "kube-api-access-n7kfz") pod "3d4a99fa-d668-4798-b988-a8d8fb9e3b28" (UID: "3d4a99fa-d668-4798-b988-a8d8fb9e3b28"). InnerVolumeSpecName "kube-api-access-n7kfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:11 crc kubenswrapper[4744]: I1003 16:43:11.815256 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d4a99fa-d668-4798-b988-a8d8fb9e3b28" (UID: "3d4a99fa-d668-4798-b988-a8d8fb9e3b28"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:11 crc kubenswrapper[4744]: I1003 16:43:11.834464 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-config-data" (OuterVolumeSpecName: "config-data") pod "3d4a99fa-d668-4798-b988-a8d8fb9e3b28" (UID: "3d4a99fa-d668-4798-b988-a8d8fb9e3b28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:11 crc kubenswrapper[4744]: I1003 16:43:11.889074 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:11 crc kubenswrapper[4744]: I1003 16:43:11.889118 4744 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:11 crc kubenswrapper[4744]: I1003 16:43:11.889132 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7kfz\" (UniqueName: \"kubernetes.io/projected/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-kube-api-access-n7kfz\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:11 crc kubenswrapper[4744]: I1003 16:43:11.889145 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d4a99fa-d668-4798-b988-a8d8fb9e3b28-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.213220 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l78wg" event={"ID":"3d4a99fa-d668-4798-b988-a8d8fb9e3b28","Type":"ContainerDied","Data":"11b4e4ac9e3f5bcfedc0ab3203c3a000d16467eadb4aaf1f4845d3f438543c69"} Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.213246 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l78wg" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.213263 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11b4e4ac9e3f5bcfedc0ab3203c3a000d16467eadb4aaf1f4845d3f438543c69" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.467687 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cpdb5" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.498943 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-combined-ca-bundle\") pod \"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe\" (UID: \"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe\") " Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.499058 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-config-data\") pod \"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe\" (UID: \"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe\") " Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.499183 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64lcw\" (UniqueName: \"kubernetes.io/projected/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-kube-api-access-64lcw\") pod \"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe\" (UID: \"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe\") " Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.504758 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-kube-api-access-64lcw" (OuterVolumeSpecName: "kube-api-access-64lcw") pod "cd5b11ee-b171-4bd8-9f40-34d9dcb428fe" (UID: "cd5b11ee-b171-4bd8-9f40-34d9dcb428fe"). InnerVolumeSpecName "kube-api-access-64lcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.532316 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd5b11ee-b171-4bd8-9f40-34d9dcb428fe" (UID: "cd5b11ee-b171-4bd8-9f40-34d9dcb428fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.603576 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.603623 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64lcw\" (UniqueName: \"kubernetes.io/projected/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-kube-api-access-64lcw\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.625823 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-config-data" (OuterVolumeSpecName: "config-data") pod "cd5b11ee-b171-4bd8-9f40-34d9dcb428fe" (UID: "cd5b11ee-b171-4bd8-9f40-34d9dcb428fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.638806 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-ds7xw"] Oct 03 16:43:12 crc kubenswrapper[4744]: E1003 16:43:12.645782 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd5b11ee-b171-4bd8-9f40-34d9dcb428fe" containerName="keystone-db-sync" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.645830 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd5b11ee-b171-4bd8-9f40-34d9dcb428fe" containerName="keystone-db-sync" Oct 03 16:43:12 crc kubenswrapper[4744]: E1003 16:43:12.645845 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97fcb340-8304-4bae-9d2b-f0b04dcea8bf" containerName="mariadb-database-create" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.645852 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="97fcb340-8304-4bae-9d2b-f0b04dcea8bf" containerName="mariadb-database-create" Oct 03 16:43:12 crc kubenswrapper[4744]: E1003 16:43:12.645888 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4a99fa-d668-4798-b988-a8d8fb9e3b28" containerName="glance-db-sync" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.645899 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4a99fa-d668-4798-b988-a8d8fb9e3b28" containerName="glance-db-sync" Oct 03 16:43:12 crc kubenswrapper[4744]: E1003 16:43:12.645939 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="299d80eb-fc43-4647-83be-a792412df7b1" containerName="mariadb-database-create" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.645946 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="299d80eb-fc43-4647-83be-a792412df7b1" containerName="mariadb-database-create" Oct 03 16:43:12 crc kubenswrapper[4744]: E1003 16:43:12.645963 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18c69bbd-4199-4da7-8e11-e05ea8680342" containerName="mariadb-database-create" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.645984 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="18c69bbd-4199-4da7-8e11-e05ea8680342" containerName="mariadb-database-create" Oct 03 16:43:12 crc kubenswrapper[4744]: E1003 16:43:12.645995 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01c5ef1-aeea-431e-a744-2be2b1a28acf" containerName="init" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.646002 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01c5ef1-aeea-431e-a744-2be2b1a28acf" containerName="init" Oct 03 16:43:12 crc kubenswrapper[4744]: E1003 16:43:12.646013 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01c5ef1-aeea-431e-a744-2be2b1a28acf" containerName="dnsmasq-dns" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.646021 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01c5ef1-aeea-431e-a744-2be2b1a28acf" containerName="dnsmasq-dns" Oct 03 16:43:12 crc kubenswrapper[4744]: E1003 16:43:12.646033 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019c5b16-937c-45d2-9003-503f4d88995d" containerName="mariadb-database-create" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.646040 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="019c5b16-937c-45d2-9003-503f4d88995d" containerName="mariadb-database-create" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.646390 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d4a99fa-d668-4798-b988-a8d8fb9e3b28" containerName="glance-db-sync" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.646402 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="18c69bbd-4199-4da7-8e11-e05ea8680342" containerName="mariadb-database-create" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.646417 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd5b11ee-b171-4bd8-9f40-34d9dcb428fe" containerName="keystone-db-sync" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.646436 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="019c5b16-937c-45d2-9003-503f4d88995d" containerName="mariadb-database-create" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.646451 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="299d80eb-fc43-4647-83be-a792412df7b1" containerName="mariadb-database-create" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.646468 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="97fcb340-8304-4bae-9d2b-f0b04dcea8bf" containerName="mariadb-database-create" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.646480 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a01c5ef1-aeea-431e-a744-2be2b1a28acf" containerName="dnsmasq-dns" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.647628 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.661455 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-ds7xw"] Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.707090 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-config\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.707157 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.707191 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.707265 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56qzq\" (UniqueName: \"kubernetes.io/projected/28402296-9623-447c-bb27-23c158859fc6-kube-api-access-56qzq\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.707327 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.707351 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.707681 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.809139 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.809194 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.809218 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-config\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.809236 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.809260 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.809321 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56qzq\" (UniqueName: \"kubernetes.io/projected/28402296-9623-447c-bb27-23c158859fc6-kube-api-access-56qzq\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.810381 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.810663 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.810713 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-config\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.810684 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.810778 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.828043 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56qzq\" (UniqueName: \"kubernetes.io/projected/28402296-9623-447c-bb27-23c158859fc6-kube-api-access-56qzq\") pod \"dnsmasq-dns-7ff5475cc9-ds7xw\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:12 crc kubenswrapper[4744]: I1003 16:43:12.974089 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.225764 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cpdb5" event={"ID":"cd5b11ee-b171-4bd8-9f40-34d9dcb428fe","Type":"ContainerDied","Data":"93e44c57540d1613f4a2b39ed0895239e15a08ab2d9beadf4526398a8ff48418"} Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.226072 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93e44c57540d1613f4a2b39ed0895239e15a08ab2d9beadf4526398a8ff48418" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.226171 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cpdb5" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.464197 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-ds7xw"] Oct 03 16:43:13 crc kubenswrapper[4744]: W1003 16:43:13.478611 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28402296_9623_447c_bb27_23c158859fc6.slice/crio-eec351c56a1c4445f864799b03b5bb929475d33ee60c547cedd7dbf87d7de3e9 WatchSource:0}: Error finding container eec351c56a1c4445f864799b03b5bb929475d33ee60c547cedd7dbf87d7de3e9: Status 404 returned error can't find the container with id eec351c56a1c4445f864799b03b5bb929475d33ee60c547cedd7dbf87d7de3e9 Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.504096 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-ds7xw"] Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.535686 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-csk6k"] Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.536706 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.563777 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.570132 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2z2zg" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.570418 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.570572 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.608159 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-csk6k"] Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.613929 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj"] Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.615290 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.627672 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-scripts\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.627738 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p54np\" (UniqueName: \"kubernetes.io/projected/4eeda15e-12ae-4c7c-ab47-669638562621-kube-api-access-p54np\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.627798 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-fernet-keys\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.627824 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-combined-ca-bundle\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.627883 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-credential-keys\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.627933 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-config-data\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.688142 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj"] Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.729809 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-config-data\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.729929 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-config\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.729975 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.730044 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-scripts\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.730068 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.730122 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p54np\" (UniqueName: \"kubernetes.io/projected/4eeda15e-12ae-4c7c-ab47-669638562621-kube-api-access-p54np\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.730143 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.730186 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-fernet-keys\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.730214 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-combined-ca-bundle\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.730274 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.730298 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzhjg\" (UniqueName: \"kubernetes.io/projected/e71535fd-11e3-4397-8af2-54c5ddf56d09-kube-api-access-nzhjg\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.730354 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-credential-keys\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.740670 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-credential-keys\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.745032 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-fernet-keys\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.745257 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-config-data\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.745578 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-scripts\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.757354 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-combined-ca-bundle\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.790133 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p54np\" (UniqueName: \"kubernetes.io/projected/4eeda15e-12ae-4c7c-ab47-669638562621-kube-api-access-p54np\") pod \"keystone-bootstrap-csk6k\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.832806 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-config\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.832864 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.832888 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.832917 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.832954 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.832971 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzhjg\" (UniqueName: \"kubernetes.io/projected/e71535fd-11e3-4397-8af2-54c5ddf56d09-kube-api-access-nzhjg\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.834213 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-config\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.834625 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.834819 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.835321 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.835914 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.865826 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.892516 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzhjg\" (UniqueName: \"kubernetes.io/projected/e71535fd-11e3-4397-8af2-54c5ddf56d09-kube-api-access-nzhjg\") pod \"dnsmasq-dns-5c5cc7c5ff-dd7nj\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.938098 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-79c99f65c-jjfnc"] Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.939317 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.977244 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.977606 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.977760 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-2w526" Oct 03 16:43:13 crc kubenswrapper[4744]: I1003 16:43:13.982288 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.018605 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79c99f65c-jjfnc"] Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.036686 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ecb56805-c2cc-4736-b5bc-84440acb7fa6-config-data\") pod \"horizon-79c99f65c-jjfnc\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.036727 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecb56805-c2cc-4736-b5bc-84440acb7fa6-logs\") pod \"horizon-79c99f65c-jjfnc\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.036751 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecb56805-c2cc-4736-b5bc-84440acb7fa6-scripts\") pod \"horizon-79c99f65c-jjfnc\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.036777 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ecb56805-c2cc-4736-b5bc-84440acb7fa6-horizon-secret-key\") pod \"horizon-79c99f65c-jjfnc\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.036810 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcpzb\" (UniqueName: \"kubernetes.io/projected/ecb56805-c2cc-4736-b5bc-84440acb7fa6-kube-api-access-mcpzb\") pod \"horizon-79c99f65c-jjfnc\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.114090 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj"] Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.115269 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.143459 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ecb56805-c2cc-4736-b5bc-84440acb7fa6-config-data\") pod \"horizon-79c99f65c-jjfnc\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.143531 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecb56805-c2cc-4736-b5bc-84440acb7fa6-logs\") pod \"horizon-79c99f65c-jjfnc\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.143584 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecb56805-c2cc-4736-b5bc-84440acb7fa6-scripts\") pod \"horizon-79c99f65c-jjfnc\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.143628 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ecb56805-c2cc-4736-b5bc-84440acb7fa6-horizon-secret-key\") pod \"horizon-79c99f65c-jjfnc\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.143674 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcpzb\" (UniqueName: \"kubernetes.io/projected/ecb56805-c2cc-4736-b5bc-84440acb7fa6-kube-api-access-mcpzb\") pod \"horizon-79c99f65c-jjfnc\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.145646 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ecb56805-c2cc-4736-b5bc-84440acb7fa6-config-data\") pod \"horizon-79c99f65c-jjfnc\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.146077 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecb56805-c2cc-4736-b5bc-84440acb7fa6-logs\") pod \"horizon-79c99f65c-jjfnc\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.146352 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecb56805-c2cc-4736-b5bc-84440acb7fa6-scripts\") pod \"horizon-79c99f65c-jjfnc\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.160319 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.161062 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ecb56805-c2cc-4736-b5bc-84440acb7fa6-horizon-secret-key\") pod \"horizon-79c99f65c-jjfnc\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.166606 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.180537 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.180898 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.180990 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-ndwh8"] Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.191602 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.198373 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcpzb\" (UniqueName: \"kubernetes.io/projected/ecb56805-c2cc-4736-b5bc-84440acb7fa6-kube-api-access-mcpzb\") pod \"horizon-79c99f65c-jjfnc\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.201448 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.202979 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.206081 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.206298 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.206426 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-fk4vm" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.206526 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.239276 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.245969 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.246011 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-scripts\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.246030 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x9h4\" (UniqueName: \"kubernetes.io/projected/7ca5eb4d-d2b1-4357-8b49-001763e93804-kube-api-access-4x9h4\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.246049 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-run-httpd\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.246066 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.246088 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.250826 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-config-data\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.250890 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.250929 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.250997 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-config\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.251020 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.251365 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-qbz46"] Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.252041 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.252120 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqc65\" (UniqueName: \"kubernetes.io/projected/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-kube-api-access-qqc65\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.252183 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-config-data\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.252200 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-scripts\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.252248 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2thc\" (UniqueName: \"kubernetes.io/projected/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-kube-api-access-c2thc\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.252280 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-logs\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.252326 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.252352 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-log-httpd\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.252376 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-ceph\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.252442 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.254200 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" event={"ID":"28402296-9623-447c-bb27-23c158859fc6","Type":"ContainerStarted","Data":"2ebddfdfb68aecf0b79d3a4ccca23a7a4867f3727b781254a45956ee4a38365a"} Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.254234 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" event={"ID":"28402296-9623-447c-bb27-23c158859fc6","Type":"ContainerStarted","Data":"eec351c56a1c4445f864799b03b5bb929475d33ee60c547cedd7dbf87d7de3e9"} Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.254331 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" podUID="28402296-9623-447c-bb27-23c158859fc6" containerName="init" containerID="cri-o://2ebddfdfb68aecf0b79d3a4ccca23a7a4867f3727b781254a45956ee4a38365a" gracePeriod=10 Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.260315 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-ndwh8"] Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.267732 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.277062 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.277752 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.277805 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-749k2" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.289139 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qbz46"] Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.309944 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.318096 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-dc6d57fbc-4s82v"] Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.320780 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.335875 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.367676 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-dc6d57fbc-4s82v"] Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.416087 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-config-data\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.416170 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-horizon-secret-key\") pod \"horizon-dc6d57fbc-4s82v\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.416201 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.416225 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z6qc\" (UniqueName: \"kubernetes.io/projected/7acfc27b-28b7-422c-a933-8dd2a4a99c68-kube-api-access-5z6qc\") pod \"placement-db-sync-qbz46\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.416262 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.416284 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7acfc27b-28b7-422c-a933-8dd2a4a99c68-logs\") pod \"placement-db-sync-qbz46\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.416343 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-config\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.416366 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.416440 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.417610 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.417810 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.422598 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-config\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.424818 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqc65\" (UniqueName: \"kubernetes.io/projected/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-kube-api-access-qqc65\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.424873 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-scripts\") pod \"horizon-dc6d57fbc-4s82v\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.424914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-config-data\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.424934 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-scripts\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.424958 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-scripts\") pod \"placement-db-sync-qbz46\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.424991 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2thc\" (UniqueName: \"kubernetes.io/projected/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-kube-api-access-c2thc\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425031 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-logs\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425064 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-config-data\") pod \"horizon-dc6d57fbc-4s82v\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425089 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425115 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-log-httpd\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425142 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-logs\") pod \"horizon-dc6d57fbc-4s82v\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425163 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-ceph\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425207 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425257 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425280 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-combined-ca-bundle\") pod \"placement-db-sync-qbz46\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425310 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b9xr\" (UniqueName: \"kubernetes.io/projected/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-kube-api-access-9b9xr\") pod \"horizon-dc6d57fbc-4s82v\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425345 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-scripts\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425372 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x9h4\" (UniqueName: \"kubernetes.io/projected/7ca5eb4d-d2b1-4357-8b49-001763e93804-kube-api-access-4x9h4\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425393 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-run-httpd\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425419 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425439 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-config-data\") pod \"placement-db-sync-qbz46\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.425472 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.428370 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-config-data\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.429645 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-logs\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.430298 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.430415 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.431306 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.432559 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.432815 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.433021 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-run-httpd\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.433245 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-log-httpd\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.435132 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-scripts\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.435665 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-scripts\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.439517 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-config-data\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.440323 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-ceph\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.467452 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.477218 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2thc\" (UniqueName: \"kubernetes.io/projected/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-kube-api-access-c2thc\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.477484 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqc65\" (UniqueName: \"kubernetes.io/projected/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-kube-api-access-qqc65\") pod \"ceilometer-0\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.478106 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x9h4\" (UniqueName: \"kubernetes.io/projected/7ca5eb4d-d2b1-4357-8b49-001763e93804-kube-api-access-4x9h4\") pod \"dnsmasq-dns-8b5c85b87-ndwh8\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.479140 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.491025 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.492916 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.494938 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.505987 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.508004 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.530228 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.531229 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7acfc27b-28b7-422c-a933-8dd2a4a99c68-logs\") pod \"placement-db-sync-qbz46\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.531356 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-scripts\") pod \"horizon-dc6d57fbc-4s82v\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.531394 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-scripts\") pod \"placement-db-sync-qbz46\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.531431 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-config-data\") pod \"horizon-dc6d57fbc-4s82v\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.531458 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-logs\") pod \"horizon-dc6d57fbc-4s82v\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.531526 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-combined-ca-bundle\") pod \"placement-db-sync-qbz46\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.531552 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b9xr\" (UniqueName: \"kubernetes.io/projected/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-kube-api-access-9b9xr\") pod \"horizon-dc6d57fbc-4s82v\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.531585 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-config-data\") pod \"placement-db-sync-qbz46\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.531638 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-horizon-secret-key\") pod \"horizon-dc6d57fbc-4s82v\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.531665 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z6qc\" (UniqueName: \"kubernetes.io/projected/7acfc27b-28b7-422c-a933-8dd2a4a99c68-kube-api-access-5z6qc\") pod \"placement-db-sync-qbz46\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.531751 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7acfc27b-28b7-422c-a933-8dd2a4a99c68-logs\") pod \"placement-db-sync-qbz46\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.532895 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-config-data\") pod \"horizon-dc6d57fbc-4s82v\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.533303 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-scripts\") pod \"horizon-dc6d57fbc-4s82v\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.533534 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-logs\") pod \"horizon-dc6d57fbc-4s82v\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.536613 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-config-data\") pod \"placement-db-sync-qbz46\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.538657 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-combined-ca-bundle\") pod \"placement-db-sync-qbz46\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.540579 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-horizon-secret-key\") pod \"horizon-dc6d57fbc-4s82v\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.548328 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.553013 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-scripts\") pod \"placement-db-sync-qbz46\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.554161 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z6qc\" (UniqueName: \"kubernetes.io/projected/7acfc27b-28b7-422c-a933-8dd2a4a99c68-kube-api-access-5z6qc\") pod \"placement-db-sync-qbz46\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.565697 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b9xr\" (UniqueName: \"kubernetes.io/projected/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-kube-api-access-9b9xr\") pod \"horizon-dc6d57fbc-4s82v\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.578504 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.633017 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.633069 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-logs\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.633106 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.633141 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.633166 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-ceph\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.633189 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.633281 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w95p9\" (UniqueName: \"kubernetes.io/projected/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-kube-api-access-w95p9\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.633308 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.661260 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qbz46" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.676987 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.708203 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-csk6k"] Oct 03 16:43:14 crc kubenswrapper[4744]: W1003 16:43:14.709227 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4eeda15e_12ae_4c7c_ab47_669638562621.slice/crio-858cbc1446284799bf597619ad46aec968c53311c5a9196db303cd6049b90686 WatchSource:0}: Error finding container 858cbc1446284799bf597619ad46aec968c53311c5a9196db303cd6049b90686: Status 404 returned error can't find the container with id 858cbc1446284799bf597619ad46aec968c53311c5a9196db303cd6049b90686 Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.736781 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w95p9\" (UniqueName: \"kubernetes.io/projected/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-kube-api-access-w95p9\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.736816 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.736885 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.736909 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-logs\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.736933 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.736956 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.736971 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-ceph\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.736995 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.737439 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.740854 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-logs\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.741064 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.749254 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.780365 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-ceph\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.784231 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.793120 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w95p9\" (UniqueName: \"kubernetes.io/projected/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-kube-api-access-w95p9\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.803437 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.848634 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:14 crc kubenswrapper[4744]: I1003 16:43:14.993784 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj"] Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.002996 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.046464 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-dns-svc\") pod \"28402296-9623-447c-bb27-23c158859fc6\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.046546 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-ovsdbserver-nb\") pod \"28402296-9623-447c-bb27-23c158859fc6\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.046584 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-config\") pod \"28402296-9623-447c-bb27-23c158859fc6\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.046715 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-dns-swift-storage-0\") pod \"28402296-9623-447c-bb27-23c158859fc6\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.046763 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56qzq\" (UniqueName: \"kubernetes.io/projected/28402296-9623-447c-bb27-23c158859fc6-kube-api-access-56qzq\") pod \"28402296-9623-447c-bb27-23c158859fc6\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.046836 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-ovsdbserver-sb\") pod \"28402296-9623-447c-bb27-23c158859fc6\" (UID: \"28402296-9623-447c-bb27-23c158859fc6\") " Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.054709 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28402296-9623-447c-bb27-23c158859fc6-kube-api-access-56qzq" (OuterVolumeSpecName: "kube-api-access-56qzq") pod "28402296-9623-447c-bb27-23c158859fc6" (UID: "28402296-9623-447c-bb27-23c158859fc6"). InnerVolumeSpecName "kube-api-access-56qzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.089654 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "28402296-9623-447c-bb27-23c158859fc6" (UID: "28402296-9623-447c-bb27-23c158859fc6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.102198 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "28402296-9623-447c-bb27-23c158859fc6" (UID: "28402296-9623-447c-bb27-23c158859fc6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.117000 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.126180 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-79c99f65c-jjfnc"] Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.126403 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "28402296-9623-447c-bb27-23c158859fc6" (UID: "28402296-9623-447c-bb27-23c158859fc6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:15 crc kubenswrapper[4744]: W1003 16:43:15.142888 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecb56805_c2cc_4736_b5bc_84440acb7fa6.slice/crio-64b7897d9087f1f10340bf44291a6b2e447d33e43b3eaccc10bd6ae15875b449 WatchSource:0}: Error finding container 64b7897d9087f1f10340bf44291a6b2e447d33e43b3eaccc10bd6ae15875b449: Status 404 returned error can't find the container with id 64b7897d9087f1f10340bf44291a6b2e447d33e43b3eaccc10bd6ae15875b449 Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.147620 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "28402296-9623-447c-bb27-23c158859fc6" (UID: "28402296-9623-447c-bb27-23c158859fc6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.149696 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.149753 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.149769 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.149778 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56qzq\" (UniqueName: \"kubernetes.io/projected/28402296-9623-447c-bb27-23c158859fc6-kube-api-access-56qzq\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.149787 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.165875 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-config" (OuterVolumeSpecName: "config") pod "28402296-9623-447c-bb27-23c158859fc6" (UID: "28402296-9623-447c-bb27-23c158859fc6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.251354 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28402296-9623-447c-bb27-23c158859fc6-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.295524 4744 generic.go:334] "Generic (PLEG): container finished" podID="28402296-9623-447c-bb27-23c158859fc6" containerID="2ebddfdfb68aecf0b79d3a4ccca23a7a4867f3727b781254a45956ee4a38365a" exitCode=0 Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.295601 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.295608 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" event={"ID":"28402296-9623-447c-bb27-23c158859fc6","Type":"ContainerDied","Data":"2ebddfdfb68aecf0b79d3a4ccca23a7a4867f3727b781254a45956ee4a38365a"} Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.295723 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-ds7xw" event={"ID":"28402296-9623-447c-bb27-23c158859fc6","Type":"ContainerDied","Data":"eec351c56a1c4445f864799b03b5bb929475d33ee60c547cedd7dbf87d7de3e9"} Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.295747 4744 scope.go:117] "RemoveContainer" containerID="2ebddfdfb68aecf0b79d3a4ccca23a7a4867f3727b781254a45956ee4a38365a" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.297122 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" event={"ID":"e71535fd-11e3-4397-8af2-54c5ddf56d09","Type":"ContainerStarted","Data":"ad4c170d62e257cfedbcd96c566e2aa8dbbcfd8a35f15ad492f380087da1ae0f"} Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.301080 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-csk6k" event={"ID":"4eeda15e-12ae-4c7c-ab47-669638562621","Type":"ContainerStarted","Data":"363010deed7d9b35da3eedc23cf2ced9cec77ca56f9bcfd98c2f8a957a5a6b26"} Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.301109 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-csk6k" event={"ID":"4eeda15e-12ae-4c7c-ab47-669638562621","Type":"ContainerStarted","Data":"858cbc1446284799bf597619ad46aec968c53311c5a9196db303cd6049b90686"} Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.305846 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79c99f65c-jjfnc" event={"ID":"ecb56805-c2cc-4736-b5bc-84440acb7fa6","Type":"ContainerStarted","Data":"64b7897d9087f1f10340bf44291a6b2e447d33e43b3eaccc10bd6ae15875b449"} Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.330457 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-csk6k" podStartSLOduration=2.330433966 podStartE2EDuration="2.330433966s" podCreationTimestamp="2025-10-03 16:43:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:15.326865925 +0000 UTC m=+1121.606741831" watchObservedRunningTime="2025-10-03 16:43:15.330433966 +0000 UTC m=+1121.610309862" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.332475 4744 scope.go:117] "RemoveContainer" containerID="2ebddfdfb68aecf0b79d3a4ccca23a7a4867f3727b781254a45956ee4a38365a" Oct 03 16:43:15 crc kubenswrapper[4744]: E1003 16:43:15.332903 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ebddfdfb68aecf0b79d3a4ccca23a7a4867f3727b781254a45956ee4a38365a\": container with ID starting with 2ebddfdfb68aecf0b79d3a4ccca23a7a4867f3727b781254a45956ee4a38365a not found: ID does not exist" containerID="2ebddfdfb68aecf0b79d3a4ccca23a7a4867f3727b781254a45956ee4a38365a" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.332935 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ebddfdfb68aecf0b79d3a4ccca23a7a4867f3727b781254a45956ee4a38365a"} err="failed to get container status \"2ebddfdfb68aecf0b79d3a4ccca23a7a4867f3727b781254a45956ee4a38365a\": rpc error: code = NotFound desc = could not find container \"2ebddfdfb68aecf0b79d3a4ccca23a7a4867f3727b781254a45956ee4a38365a\": container with ID starting with 2ebddfdfb68aecf0b79d3a4ccca23a7a4867f3727b781254a45956ee4a38365a not found: ID does not exist" Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.365785 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qbz46"] Oct 03 16:43:15 crc kubenswrapper[4744]: W1003 16:43:15.385269 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7acfc27b_28b7_422c_a933_8dd2a4a99c68.slice/crio-28e3444a85024e03ece9face876aab8a91e8b2967b4a712f64ce863e5052661b WatchSource:0}: Error finding container 28e3444a85024e03ece9face876aab8a91e8b2967b4a712f64ce863e5052661b: Status 404 returned error can't find the container with id 28e3444a85024e03ece9face876aab8a91e8b2967b4a712f64ce863e5052661b Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.418802 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-ds7xw"] Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.433701 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-ds7xw"] Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.443699 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-ndwh8"] Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.452803 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:43:15 crc kubenswrapper[4744]: I1003 16:43:15.463849 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-dc6d57fbc-4s82v"] Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:15.476951 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 16:43:16 crc kubenswrapper[4744]: W1003 16:43:15.498725 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ea987f3_4a2b_4f52_a5ef_104dc793f2fc.slice/crio-be311d43401ab3a383990ec423ff8c5ea94f49668d7cd94fab5fcd57937b3af6 WatchSource:0}: Error finding container be311d43401ab3a383990ec423ff8c5ea94f49668d7cd94fab5fcd57937b3af6: Status 404 returned error can't find the container with id be311d43401ab3a383990ec423ff8c5ea94f49668d7cd94fab5fcd57937b3af6 Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:15.767773 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.093849 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-0d09-account-create-g4bpt"] Oct 03 16:43:16 crc kubenswrapper[4744]: E1003 16:43:16.094515 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28402296-9623-447c-bb27-23c158859fc6" containerName="init" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.094532 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="28402296-9623-447c-bb27-23c158859fc6" containerName="init" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.094753 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="28402296-9623-447c-bb27-23c158859fc6" containerName="init" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.095290 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0d09-account-create-g4bpt" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.102625 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.130574 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0d09-account-create-g4bpt"] Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.189463 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqfmt\" (UniqueName: \"kubernetes.io/projected/aa76346f-0ca5-4bad-8fb8-802077202e39-kube-api-access-rqfmt\") pod \"barbican-0d09-account-create-g4bpt\" (UID: \"aa76346f-0ca5-4bad-8fb8-802077202e39\") " pod="openstack/barbican-0d09-account-create-g4bpt" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.292981 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqfmt\" (UniqueName: \"kubernetes.io/projected/aa76346f-0ca5-4bad-8fb8-802077202e39-kube-api-access-rqfmt\") pod \"barbican-0d09-account-create-g4bpt\" (UID: \"aa76346f-0ca5-4bad-8fb8-802077202e39\") " pod="openstack/barbican-0d09-account-create-g4bpt" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.322595 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-ffe1-account-create-nkjkf"] Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.324002 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ffe1-account-create-nkjkf" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.325691 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqfmt\" (UniqueName: \"kubernetes.io/projected/aa76346f-0ca5-4bad-8fb8-802077202e39-kube-api-access-rqfmt\") pod \"barbican-0d09-account-create-g4bpt\" (UID: \"aa76346f-0ca5-4bad-8fb8-802077202e39\") " pod="openstack/barbican-0d09-account-create-g4bpt" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.327394 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.352025 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-ffe1-account-create-nkjkf"] Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.373992 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc","Type":"ContainerStarted","Data":"be311d43401ab3a383990ec423ff8c5ea94f49668d7cd94fab5fcd57937b3af6"} Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.383155 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qbz46" event={"ID":"7acfc27b-28b7-422c-a933-8dd2a4a99c68","Type":"ContainerStarted","Data":"28e3444a85024e03ece9face876aab8a91e8b2967b4a712f64ce863e5052661b"} Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.385623 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-3d6b-account-create-klnll"] Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.387262 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3d6b-account-create-klnll" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.389765 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.394576 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3d6b-account-create-klnll"] Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.394985 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9pn5\" (UniqueName: \"kubernetes.io/projected/c4594c72-66b3-410f-b2ef-dbb205807338-kube-api-access-v9pn5\") pod \"manila-ffe1-account-create-nkjkf\" (UID: \"c4594c72-66b3-410f-b2ef-dbb205807338\") " pod="openstack/manila-ffe1-account-create-nkjkf" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.408350 4744 generic.go:334] "Generic (PLEG): container finished" podID="e71535fd-11e3-4397-8af2-54c5ddf56d09" containerID="d39a8a9daa99f5af871278dc9136330a814efb8b52bb5f10a194304af3d9e092" exitCode=0 Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.408441 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" event={"ID":"e71535fd-11e3-4397-8af2-54c5ddf56d09","Type":"ContainerDied","Data":"d39a8a9daa99f5af871278dc9136330a814efb8b52bb5f10a194304af3d9e092"} Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.426887 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4","Type":"ContainerStarted","Data":"a514666c722adef5d24f29fe37576007ea0b02af93a2eb43e825f763ced6e1eb"} Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.438586 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dc6d57fbc-4s82v" event={"ID":"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc","Type":"ContainerStarted","Data":"e6299623c2bfc959a73601967bea901378c7bc147dc129ded91ff9aae570dae1"} Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.453204 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1d2b85de-e61a-4a88-8180-c0cc398b7e6d","Type":"ContainerStarted","Data":"3df562a2b97659cf0c6868538a7fe8b42b6f16622344768ac77ea0a67e184c97"} Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.473285 4744 generic.go:334] "Generic (PLEG): container finished" podID="7ca5eb4d-d2b1-4357-8b49-001763e93804" containerID="f371500e59388e92174a2cefe793cdff8c9dc984766822aa684c5ba02fa610c3" exitCode=0 Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.473367 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" event={"ID":"7ca5eb4d-d2b1-4357-8b49-001763e93804","Type":"ContainerDied","Data":"f371500e59388e92174a2cefe793cdff8c9dc984766822aa684c5ba02fa610c3"} Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.473395 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" event={"ID":"7ca5eb4d-d2b1-4357-8b49-001763e93804","Type":"ContainerStarted","Data":"acc9300193f94ed5e5b824c1a800330440d35a65ad1448f4fb20df9d1fce861d"} Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.480773 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0d09-account-create-g4bpt" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.496885 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9pn5\" (UniqueName: \"kubernetes.io/projected/c4594c72-66b3-410f-b2ef-dbb205807338-kube-api-access-v9pn5\") pod \"manila-ffe1-account-create-nkjkf\" (UID: \"c4594c72-66b3-410f-b2ef-dbb205807338\") " pod="openstack/manila-ffe1-account-create-nkjkf" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.497009 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pp92\" (UniqueName: \"kubernetes.io/projected/9be7c905-97b5-40a2-b416-ff8894216fb4-kube-api-access-4pp92\") pod \"cinder-3d6b-account-create-klnll\" (UID: \"9be7c905-97b5-40a2-b416-ff8894216fb4\") " pod="openstack/cinder-3d6b-account-create-klnll" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.517909 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9pn5\" (UniqueName: \"kubernetes.io/projected/c4594c72-66b3-410f-b2ef-dbb205807338-kube-api-access-v9pn5\") pod \"manila-ffe1-account-create-nkjkf\" (UID: \"c4594c72-66b3-410f-b2ef-dbb205807338\") " pod="openstack/manila-ffe1-account-create-nkjkf" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.599232 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pp92\" (UniqueName: \"kubernetes.io/projected/9be7c905-97b5-40a2-b416-ff8894216fb4-kube-api-access-4pp92\") pod \"cinder-3d6b-account-create-klnll\" (UID: \"9be7c905-97b5-40a2-b416-ff8894216fb4\") " pod="openstack/cinder-3d6b-account-create-klnll" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.617407 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pp92\" (UniqueName: \"kubernetes.io/projected/9be7c905-97b5-40a2-b416-ff8894216fb4-kube-api-access-4pp92\") pod \"cinder-3d6b-account-create-klnll\" (UID: \"9be7c905-97b5-40a2-b416-ff8894216fb4\") " pod="openstack/cinder-3d6b-account-create-klnll" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.684526 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-2091-account-create-wz79h"] Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.686157 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2091-account-create-wz79h" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.689360 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.717110 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2091-account-create-wz79h"] Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.743071 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ffe1-account-create-nkjkf" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.760941 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3d6b-account-create-klnll" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.803075 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sxfg\" (UniqueName: \"kubernetes.io/projected/08780af0-0f6f-419c-8836-2e4661c5ff70-kube-api-access-8sxfg\") pod \"neutron-2091-account-create-wz79h\" (UID: \"08780af0-0f6f-419c-8836-2e4661c5ff70\") " pod="openstack/neutron-2091-account-create-wz79h" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.905651 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sxfg\" (UniqueName: \"kubernetes.io/projected/08780af0-0f6f-419c-8836-2e4661c5ff70-kube-api-access-8sxfg\") pod \"neutron-2091-account-create-wz79h\" (UID: \"08780af0-0f6f-419c-8836-2e4661c5ff70\") " pod="openstack/neutron-2091-account-create-wz79h" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.935131 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sxfg\" (UniqueName: \"kubernetes.io/projected/08780af0-0f6f-419c-8836-2e4661c5ff70-kube-api-access-8sxfg\") pod \"neutron-2091-account-create-wz79h\" (UID: \"08780af0-0f6f-419c-8836-2e4661c5ff70\") " pod="openstack/neutron-2091-account-create-wz79h" Oct 03 16:43:16 crc kubenswrapper[4744]: I1003 16:43:16.949865 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28402296-9623-447c-bb27-23c158859fc6" path="/var/lib/kubelet/pods/28402296-9623-447c-bb27-23c158859fc6/volumes" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.016953 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2091-account-create-wz79h" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.054615 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.111362 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzhjg\" (UniqueName: \"kubernetes.io/projected/e71535fd-11e3-4397-8af2-54c5ddf56d09-kube-api-access-nzhjg\") pod \"e71535fd-11e3-4397-8af2-54c5ddf56d09\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.111556 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-ovsdbserver-nb\") pod \"e71535fd-11e3-4397-8af2-54c5ddf56d09\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.111641 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-ovsdbserver-sb\") pod \"e71535fd-11e3-4397-8af2-54c5ddf56d09\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.111697 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-dns-swift-storage-0\") pod \"e71535fd-11e3-4397-8af2-54c5ddf56d09\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.111740 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-config\") pod \"e71535fd-11e3-4397-8af2-54c5ddf56d09\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.111786 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-dns-svc\") pod \"e71535fd-11e3-4397-8af2-54c5ddf56d09\" (UID: \"e71535fd-11e3-4397-8af2-54c5ddf56d09\") " Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.125762 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e71535fd-11e3-4397-8af2-54c5ddf56d09-kube-api-access-nzhjg" (OuterVolumeSpecName: "kube-api-access-nzhjg") pod "e71535fd-11e3-4397-8af2-54c5ddf56d09" (UID: "e71535fd-11e3-4397-8af2-54c5ddf56d09"). InnerVolumeSpecName "kube-api-access-nzhjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.150187 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e71535fd-11e3-4397-8af2-54c5ddf56d09" (UID: "e71535fd-11e3-4397-8af2-54c5ddf56d09"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.150274 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-config" (OuterVolumeSpecName: "config") pod "e71535fd-11e3-4397-8af2-54c5ddf56d09" (UID: "e71535fd-11e3-4397-8af2-54c5ddf56d09"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.162860 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e71535fd-11e3-4397-8af2-54c5ddf56d09" (UID: "e71535fd-11e3-4397-8af2-54c5ddf56d09"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.167419 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e71535fd-11e3-4397-8af2-54c5ddf56d09" (UID: "e71535fd-11e3-4397-8af2-54c5ddf56d09"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.171913 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0d09-account-create-g4bpt"] Oct 03 16:43:17 crc kubenswrapper[4744]: W1003 16:43:17.181834 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa76346f_0ca5_4bad_8fb8_802077202e39.slice/crio-a8fafd8891a38659d697bf71c1752465b261832828bb820f414d322d0c9b5fb4 WatchSource:0}: Error finding container a8fafd8891a38659d697bf71c1752465b261832828bb820f414d322d0c9b5fb4: Status 404 returned error can't find the container with id a8fafd8891a38659d697bf71c1752465b261832828bb820f414d322d0c9b5fb4 Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.186677 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e71535fd-11e3-4397-8af2-54c5ddf56d09" (UID: "e71535fd-11e3-4397-8af2-54c5ddf56d09"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.214221 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.214247 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.214258 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.214266 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.214274 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzhjg\" (UniqueName: \"kubernetes.io/projected/e71535fd-11e3-4397-8af2-54c5ddf56d09-kube-api-access-nzhjg\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.214286 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e71535fd-11e3-4397-8af2-54c5ddf56d09-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.493971 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc","Type":"ContainerStarted","Data":"ab06515926dcd29c8cf7aebb1c3a15cfa8fc923a716d85ca72850c8e24339a23"} Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.496348 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" event={"ID":"e71535fd-11e3-4397-8af2-54c5ddf56d09","Type":"ContainerDied","Data":"ad4c170d62e257cfedbcd96c566e2aa8dbbcfd8a35f15ad492f380087da1ae0f"} Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.496393 4744 scope.go:117] "RemoveContainer" containerID="d39a8a9daa99f5af871278dc9136330a814efb8b52bb5f10a194304af3d9e092" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.498145 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.506275 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0d09-account-create-g4bpt" event={"ID":"aa76346f-0ca5-4bad-8fb8-802077202e39","Type":"ContainerStarted","Data":"a8fafd8891a38659d697bf71c1752465b261832828bb820f414d322d0c9b5fb4"} Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.514808 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1d2b85de-e61a-4a88-8180-c0cc398b7e6d","Type":"ContainerStarted","Data":"a05cc65f56c3d394943b07d758b31ef01f67118f473ee2b9c3bb51af31214e3e"} Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.526162 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" event={"ID":"7ca5eb4d-d2b1-4357-8b49-001763e93804","Type":"ContainerStarted","Data":"42af22cf2a27fc349e7615656c18eb71e7b78bf252206da91e54c7bcca85a59a"} Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.526341 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.558774 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" podStartSLOduration=3.558749778 podStartE2EDuration="3.558749778s" podCreationTimestamp="2025-10-03 16:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:17.55175399 +0000 UTC m=+1123.831629886" watchObservedRunningTime="2025-10-03 16:43:17.558749778 +0000 UTC m=+1123.838625674" Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.579014 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-ffe1-account-create-nkjkf"] Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.624163 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj"] Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.633593 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-dd7nj"] Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.655205 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3d6b-account-create-klnll"] Oct 03 16:43:17 crc kubenswrapper[4744]: I1003 16:43:17.676471 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2091-account-create-wz79h"] Oct 03 16:43:18 crc kubenswrapper[4744]: I1003 16:43:18.535220 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0d09-account-create-g4bpt" event={"ID":"aa76346f-0ca5-4bad-8fb8-802077202e39","Type":"ContainerStarted","Data":"7b951a5e87e97262961e2d9245e35801cb16b1e333e328900801c6e43581e69b"} Oct 03 16:43:18 crc kubenswrapper[4744]: I1003 16:43:18.539190 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2091-account-create-wz79h" event={"ID":"08780af0-0f6f-419c-8836-2e4661c5ff70","Type":"ContainerStarted","Data":"5b45fe31d0bbc237269c387a96fc87e1a790437b80f084fa72f5b790cf2aa152"} Oct 03 16:43:18 crc kubenswrapper[4744]: I1003 16:43:18.541102 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3d6b-account-create-klnll" event={"ID":"9be7c905-97b5-40a2-b416-ff8894216fb4","Type":"ContainerStarted","Data":"df72c96120b22a6825e817ce75f8450897fc4f9fbda38e35d5fbfb7224a5544a"} Oct 03 16:43:18 crc kubenswrapper[4744]: I1003 16:43:18.544061 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-ffe1-account-create-nkjkf" event={"ID":"c4594c72-66b3-410f-b2ef-dbb205807338","Type":"ContainerStarted","Data":"9d642b49d4045879438badbedb672b316c8255f4b024649ce76b3aecb889e76d"} Oct 03 16:43:18 crc kubenswrapper[4744]: I1003 16:43:18.553559 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-0d09-account-create-g4bpt" podStartSLOduration=2.55354172 podStartE2EDuration="2.55354172s" podCreationTimestamp="2025-10-03 16:43:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:18.549353745 +0000 UTC m=+1124.829229641" watchObservedRunningTime="2025-10-03 16:43:18.55354172 +0000 UTC m=+1124.833417616" Oct 03 16:43:18 crc kubenswrapper[4744]: I1003 16:43:18.852178 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 16:43:18 crc kubenswrapper[4744]: I1003 16:43:18.911643 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e71535fd-11e3-4397-8af2-54c5ddf56d09" path="/var/lib/kubelet/pods/e71535fd-11e3-4397-8af2-54c5ddf56d09/volumes" Oct 03 16:43:18 crc kubenswrapper[4744]: I1003 16:43:18.912174 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:43:18 crc kubenswrapper[4744]: I1003 16:43:18.955106 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79c99f65c-jjfnc"] Oct 03 16:43:18 crc kubenswrapper[4744]: I1003 16:43:18.984240 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.026228 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-75bcc4d5c-s79cg"] Oct 03 16:43:19 crc kubenswrapper[4744]: E1003 16:43:19.026662 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e71535fd-11e3-4397-8af2-54c5ddf56d09" containerName="init" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.026678 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e71535fd-11e3-4397-8af2-54c5ddf56d09" containerName="init" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.026848 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e71535fd-11e3-4397-8af2-54c5ddf56d09" containerName="init" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.027751 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.034922 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-75bcc4d5c-s79cg"] Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.174354 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d991f026-62e5-4515-924d-ae7267cab11f-scripts\") pod \"horizon-75bcc4d5c-s79cg\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.174515 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d991f026-62e5-4515-924d-ae7267cab11f-horizon-secret-key\") pod \"horizon-75bcc4d5c-s79cg\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.174555 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d991f026-62e5-4515-924d-ae7267cab11f-logs\") pod \"horizon-75bcc4d5c-s79cg\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.174592 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmxsd\" (UniqueName: \"kubernetes.io/projected/d991f026-62e5-4515-924d-ae7267cab11f-kube-api-access-zmxsd\") pod \"horizon-75bcc4d5c-s79cg\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.174619 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d991f026-62e5-4515-924d-ae7267cab11f-config-data\") pod \"horizon-75bcc4d5c-s79cg\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.276184 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d991f026-62e5-4515-924d-ae7267cab11f-scripts\") pod \"horizon-75bcc4d5c-s79cg\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.276297 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d991f026-62e5-4515-924d-ae7267cab11f-horizon-secret-key\") pod \"horizon-75bcc4d5c-s79cg\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.276360 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d991f026-62e5-4515-924d-ae7267cab11f-logs\") pod \"horizon-75bcc4d5c-s79cg\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.276391 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmxsd\" (UniqueName: \"kubernetes.io/projected/d991f026-62e5-4515-924d-ae7267cab11f-kube-api-access-zmxsd\") pod \"horizon-75bcc4d5c-s79cg\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.276426 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d991f026-62e5-4515-924d-ae7267cab11f-config-data\") pod \"horizon-75bcc4d5c-s79cg\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.284364 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d991f026-62e5-4515-924d-ae7267cab11f-logs\") pod \"horizon-75bcc4d5c-s79cg\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.284583 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d991f026-62e5-4515-924d-ae7267cab11f-config-data\") pod \"horizon-75bcc4d5c-s79cg\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.285151 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d991f026-62e5-4515-924d-ae7267cab11f-scripts\") pod \"horizon-75bcc4d5c-s79cg\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.304284 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d991f026-62e5-4515-924d-ae7267cab11f-horizon-secret-key\") pod \"horizon-75bcc4d5c-s79cg\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.308287 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmxsd\" (UniqueName: \"kubernetes.io/projected/d991f026-62e5-4515-924d-ae7267cab11f-kube-api-access-zmxsd\") pod \"horizon-75bcc4d5c-s79cg\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.377077 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.567064 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-ffe1-account-create-nkjkf" event={"ID":"c4594c72-66b3-410f-b2ef-dbb205807338","Type":"ContainerStarted","Data":"aa607b01ba85a1bd8828075e6bd1127d39a8420d2c70e3da63381ffd39d72f89"} Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.570083 4744 generic.go:334] "Generic (PLEG): container finished" podID="aa76346f-0ca5-4bad-8fb8-802077202e39" containerID="7b951a5e87e97262961e2d9245e35801cb16b1e333e328900801c6e43581e69b" exitCode=0 Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.570140 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0d09-account-create-g4bpt" event={"ID":"aa76346f-0ca5-4bad-8fb8-802077202e39","Type":"ContainerDied","Data":"7b951a5e87e97262961e2d9245e35801cb16b1e333e328900801c6e43581e69b"} Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.573754 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1d2b85de-e61a-4a88-8180-c0cc398b7e6d","Type":"ContainerStarted","Data":"2ecdf210efa873bc6811289e1fba3768d387757a380c9253b160612617cd70ef"} Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.576290 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc","Type":"ContainerStarted","Data":"1023f99e320002af835fecc4de9f9781c1fd5de2d6dbba843abdf1fe761a501e"} Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.576618 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" containerName="glance-log" containerID="cri-o://ab06515926dcd29c8cf7aebb1c3a15cfa8fc923a716d85ca72850c8e24339a23" gracePeriod=30 Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.576645 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" containerName="glance-httpd" containerID="cri-o://1023f99e320002af835fecc4de9f9781c1fd5de2d6dbba843abdf1fe761a501e" gracePeriod=30 Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.580337 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2091-account-create-wz79h" event={"ID":"08780af0-0f6f-419c-8836-2e4661c5ff70","Type":"ContainerStarted","Data":"f505d3b8552d2d84f99a727f75a0601cc01798cdb6435537d32469da7debca61"} Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.583258 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3d6b-account-create-klnll" event={"ID":"9be7c905-97b5-40a2-b416-ff8894216fb4","Type":"ContainerStarted","Data":"786ebc5ff90c568b00f25cde96f5779e594932c2c57b36ba551e611472ffa602"} Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.602671 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-ffe1-account-create-nkjkf" podStartSLOduration=3.602651831 podStartE2EDuration="3.602651831s" podCreationTimestamp="2025-10-03 16:43:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:19.587254442 +0000 UTC m=+1125.867130338" watchObservedRunningTime="2025-10-03 16:43:19.602651831 +0000 UTC m=+1125.882527727" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.616145 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.616124716 podStartE2EDuration="5.616124716s" podCreationTimestamp="2025-10-03 16:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:19.613762843 +0000 UTC m=+1125.893638759" watchObservedRunningTime="2025-10-03 16:43:19.616124716 +0000 UTC m=+1125.896000612" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.633607 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-3d6b-account-create-klnll" podStartSLOduration=3.633592132 podStartE2EDuration="3.633592132s" podCreationTimestamp="2025-10-03 16:43:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:19.629456669 +0000 UTC m=+1125.909332565" watchObservedRunningTime="2025-10-03 16:43:19.633592132 +0000 UTC m=+1125.913468028" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.657838 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.657818182 podStartE2EDuration="5.657818182s" podCreationTimestamp="2025-10-03 16:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:19.651511819 +0000 UTC m=+1125.931387735" watchObservedRunningTime="2025-10-03 16:43:19.657818182 +0000 UTC m=+1125.937694078" Oct 03 16:43:19 crc kubenswrapper[4744]: I1003 16:43:19.671742 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-2091-account-create-wz79h" podStartSLOduration=3.6717258470000003 podStartE2EDuration="3.671725847s" podCreationTimestamp="2025-10-03 16:43:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:19.663197314 +0000 UTC m=+1125.943073210" watchObservedRunningTime="2025-10-03 16:43:19.671725847 +0000 UTC m=+1125.951601743" Oct 03 16:43:20 crc kubenswrapper[4744]: I1003 16:43:20.594964 4744 generic.go:334] "Generic (PLEG): container finished" podID="08780af0-0f6f-419c-8836-2e4661c5ff70" containerID="f505d3b8552d2d84f99a727f75a0601cc01798cdb6435537d32469da7debca61" exitCode=0 Oct 03 16:43:20 crc kubenswrapper[4744]: I1003 16:43:20.595414 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2091-account-create-wz79h" event={"ID":"08780af0-0f6f-419c-8836-2e4661c5ff70","Type":"ContainerDied","Data":"f505d3b8552d2d84f99a727f75a0601cc01798cdb6435537d32469da7debca61"} Oct 03 16:43:20 crc kubenswrapper[4744]: I1003 16:43:20.600159 4744 generic.go:334] "Generic (PLEG): container finished" podID="9be7c905-97b5-40a2-b416-ff8894216fb4" containerID="786ebc5ff90c568b00f25cde96f5779e594932c2c57b36ba551e611472ffa602" exitCode=0 Oct 03 16:43:20 crc kubenswrapper[4744]: I1003 16:43:20.600251 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3d6b-account-create-klnll" event={"ID":"9be7c905-97b5-40a2-b416-ff8894216fb4","Type":"ContainerDied","Data":"786ebc5ff90c568b00f25cde96f5779e594932c2c57b36ba551e611472ffa602"} Oct 03 16:43:20 crc kubenswrapper[4744]: I1003 16:43:20.602032 4744 generic.go:334] "Generic (PLEG): container finished" podID="c4594c72-66b3-410f-b2ef-dbb205807338" containerID="aa607b01ba85a1bd8828075e6bd1127d39a8420d2c70e3da63381ffd39d72f89" exitCode=0 Oct 03 16:43:20 crc kubenswrapper[4744]: I1003 16:43:20.602071 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-ffe1-account-create-nkjkf" event={"ID":"c4594c72-66b3-410f-b2ef-dbb205807338","Type":"ContainerDied","Data":"aa607b01ba85a1bd8828075e6bd1127d39a8420d2c70e3da63381ffd39d72f89"} Oct 03 16:43:20 crc kubenswrapper[4744]: I1003 16:43:20.605013 4744 generic.go:334] "Generic (PLEG): container finished" podID="8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" containerID="1023f99e320002af835fecc4de9f9781c1fd5de2d6dbba843abdf1fe761a501e" exitCode=0 Oct 03 16:43:20 crc kubenswrapper[4744]: I1003 16:43:20.605052 4744 generic.go:334] "Generic (PLEG): container finished" podID="8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" containerID="ab06515926dcd29c8cf7aebb1c3a15cfa8fc923a716d85ca72850c8e24339a23" exitCode=143 Oct 03 16:43:20 crc kubenswrapper[4744]: I1003 16:43:20.605131 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc","Type":"ContainerDied","Data":"1023f99e320002af835fecc4de9f9781c1fd5de2d6dbba843abdf1fe761a501e"} Oct 03 16:43:20 crc kubenswrapper[4744]: I1003 16:43:20.605158 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc","Type":"ContainerDied","Data":"ab06515926dcd29c8cf7aebb1c3a15cfa8fc923a716d85ca72850c8e24339a23"} Oct 03 16:43:20 crc kubenswrapper[4744]: I1003 16:43:20.605300 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1d2b85de-e61a-4a88-8180-c0cc398b7e6d" containerName="glance-log" containerID="cri-o://a05cc65f56c3d394943b07d758b31ef01f67118f473ee2b9c3bb51af31214e3e" gracePeriod=30 Oct 03 16:43:20 crc kubenswrapper[4744]: I1003 16:43:20.605347 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1d2b85de-e61a-4a88-8180-c0cc398b7e6d" containerName="glance-httpd" containerID="cri-o://2ecdf210efa873bc6811289e1fba3768d387757a380c9253b160612617cd70ef" gracePeriod=30 Oct 03 16:43:21 crc kubenswrapper[4744]: I1003 16:43:21.618188 4744 generic.go:334] "Generic (PLEG): container finished" podID="4eeda15e-12ae-4c7c-ab47-669638562621" containerID="363010deed7d9b35da3eedc23cf2ced9cec77ca56f9bcfd98c2f8a957a5a6b26" exitCode=0 Oct 03 16:43:21 crc kubenswrapper[4744]: I1003 16:43:21.618273 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-csk6k" event={"ID":"4eeda15e-12ae-4c7c-ab47-669638562621","Type":"ContainerDied","Data":"363010deed7d9b35da3eedc23cf2ced9cec77ca56f9bcfd98c2f8a957a5a6b26"} Oct 03 16:43:21 crc kubenswrapper[4744]: I1003 16:43:21.620468 4744 generic.go:334] "Generic (PLEG): container finished" podID="1d2b85de-e61a-4a88-8180-c0cc398b7e6d" containerID="2ecdf210efa873bc6811289e1fba3768d387757a380c9253b160612617cd70ef" exitCode=0 Oct 03 16:43:21 crc kubenswrapper[4744]: I1003 16:43:21.620510 4744 generic.go:334] "Generic (PLEG): container finished" podID="1d2b85de-e61a-4a88-8180-c0cc398b7e6d" containerID="a05cc65f56c3d394943b07d758b31ef01f67118f473ee2b9c3bb51af31214e3e" exitCode=143 Oct 03 16:43:21 crc kubenswrapper[4744]: I1003 16:43:21.620514 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1d2b85de-e61a-4a88-8180-c0cc398b7e6d","Type":"ContainerDied","Data":"2ecdf210efa873bc6811289e1fba3768d387757a380c9253b160612617cd70ef"} Oct 03 16:43:21 crc kubenswrapper[4744]: I1003 16:43:21.620548 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1d2b85de-e61a-4a88-8180-c0cc398b7e6d","Type":"ContainerDied","Data":"a05cc65f56c3d394943b07d758b31ef01f67118f473ee2b9c3bb51af31214e3e"} Oct 03 16:43:22 crc kubenswrapper[4744]: I1003 16:43:22.137838 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2091-account-create-wz79h" Oct 03 16:43:22 crc kubenswrapper[4744]: I1003 16:43:22.231628 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sxfg\" (UniqueName: \"kubernetes.io/projected/08780af0-0f6f-419c-8836-2e4661c5ff70-kube-api-access-8sxfg\") pod \"08780af0-0f6f-419c-8836-2e4661c5ff70\" (UID: \"08780af0-0f6f-419c-8836-2e4661c5ff70\") " Oct 03 16:43:22 crc kubenswrapper[4744]: I1003 16:43:22.284802 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08780af0-0f6f-419c-8836-2e4661c5ff70-kube-api-access-8sxfg" (OuterVolumeSpecName: "kube-api-access-8sxfg") pod "08780af0-0f6f-419c-8836-2e4661c5ff70" (UID: "08780af0-0f6f-419c-8836-2e4661c5ff70"). InnerVolumeSpecName "kube-api-access-8sxfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:22 crc kubenswrapper[4744]: I1003 16:43:22.333549 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sxfg\" (UniqueName: \"kubernetes.io/projected/08780af0-0f6f-419c-8836-2e4661c5ff70-kube-api-access-8sxfg\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:22 crc kubenswrapper[4744]: I1003 16:43:22.631163 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2091-account-create-wz79h" event={"ID":"08780af0-0f6f-419c-8836-2e4661c5ff70","Type":"ContainerDied","Data":"5b45fe31d0bbc237269c387a96fc87e1a790437b80f084fa72f5b790cf2aa152"} Oct 03 16:43:22 crc kubenswrapper[4744]: I1003 16:43:22.631235 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b45fe31d0bbc237269c387a96fc87e1a790437b80f084fa72f5b790cf2aa152" Oct 03 16:43:22 crc kubenswrapper[4744]: I1003 16:43:22.631180 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2091-account-create-wz79h" Oct 03 16:43:24 crc kubenswrapper[4744]: I1003 16:43:24.552948 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:43:24 crc kubenswrapper[4744]: I1003 16:43:24.637081 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-8hrkj"] Oct 03 16:43:24 crc kubenswrapper[4744]: I1003 16:43:24.637326 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" podUID="3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" containerName="dnsmasq-dns" containerID="cri-o://1228505882f8ef70a3feeb4a8aecc382e7742259e6fb684543f703f4163038a5" gracePeriod=10 Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.626286 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-dc6d57fbc-4s82v"] Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.646045 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" podUID="3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.678211 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7978488848-g4fcf"] Oct 03 16:43:25 crc kubenswrapper[4744]: E1003 16:43:25.678780 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08780af0-0f6f-419c-8836-2e4661c5ff70" containerName="mariadb-account-create" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.678796 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="08780af0-0f6f-419c-8836-2e4661c5ff70" containerName="mariadb-account-create" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.679436 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="08780af0-0f6f-419c-8836-2e4661c5ff70" containerName="mariadb-account-create" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.686260 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.694818 4744 generic.go:334] "Generic (PLEG): container finished" podID="3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" containerID="1228505882f8ef70a3feeb4a8aecc382e7742259e6fb684543f703f4163038a5" exitCode=0 Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.694872 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" event={"ID":"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914","Type":"ContainerDied","Data":"1228505882f8ef70a3feeb4a8aecc382e7742259e6fb684543f703f4163038a5"} Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.695007 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.706746 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7978488848-g4fcf"] Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.748717 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-75bcc4d5c-s79cg"] Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.776672 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-85d8c4f7b-9pn28"] Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.779111 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.794297 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85d8c4f7b-9pn28"] Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.798322 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/874834f7-8c7c-422d-9ec1-822d591f6d75-config-data\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.798378 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-combined-ca-bundle\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.798405 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/874834f7-8c7c-422d-9ec1-822d591f6d75-logs\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.798514 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drh7c\" (UniqueName: \"kubernetes.io/projected/874834f7-8c7c-422d-9ec1-822d591f6d75-kube-api-access-drh7c\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.798548 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-horizon-secret-key\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.798608 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-scripts\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.798624 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-combined-ca-bundle\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.798644 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-horizon-secret-key\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.798682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-horizon-tls-certs\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.798710 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/874834f7-8c7c-422d-9ec1-822d591f6d75-scripts\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.798728 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q6z6\" (UniqueName: \"kubernetes.io/projected/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-kube-api-access-8q6z6\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.798778 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-horizon-tls-certs\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.798799 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-config-data\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.798844 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-logs\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.900673 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-horizon-secret-key\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.901037 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-scripts\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.901133 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-combined-ca-bundle\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.901210 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-horizon-tls-certs\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.901297 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-horizon-secret-key\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.901380 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/874834f7-8c7c-422d-9ec1-822d591f6d75-scripts\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.901458 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q6z6\" (UniqueName: \"kubernetes.io/projected/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-kube-api-access-8q6z6\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.901593 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-horizon-tls-certs\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.901703 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-config-data\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.901796 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-logs\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.901928 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/874834f7-8c7c-422d-9ec1-822d591f6d75-config-data\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.902039 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-combined-ca-bundle\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.902142 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/874834f7-8c7c-422d-9ec1-822d591f6d75-logs\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.902244 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drh7c\" (UniqueName: \"kubernetes.io/projected/874834f7-8c7c-422d-9ec1-822d591f6d75-kube-api-access-drh7c\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.903021 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/874834f7-8c7c-422d-9ec1-822d591f6d75-scripts\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.905479 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/874834f7-8c7c-422d-9ec1-822d591f6d75-logs\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.905615 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-logs\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.905773 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-scripts\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.907076 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/874834f7-8c7c-422d-9ec1-822d591f6d75-config-data\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.907874 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-combined-ca-bundle\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.907988 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-horizon-secret-key\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.908804 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-horizon-tls-certs\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.909364 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-horizon-secret-key\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.910027 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-combined-ca-bundle\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.917280 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-config-data\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.920763 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-horizon-tls-certs\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.926972 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drh7c\" (UniqueName: \"kubernetes.io/projected/874834f7-8c7c-422d-9ec1-822d591f6d75-kube-api-access-drh7c\") pod \"horizon-7978488848-g4fcf\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:25 crc kubenswrapper[4744]: I1003 16:43:25.929397 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q6z6\" (UniqueName: \"kubernetes.io/projected/39c711a0-3cea-4091-94a0-ec9df6e6a2d8-kube-api-access-8q6z6\") pod \"horizon-85d8c4f7b-9pn28\" (UID: \"39c711a0-3cea-4091-94a0-ec9df6e6a2d8\") " pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:26 crc kubenswrapper[4744]: I1003 16:43:26.015123 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:26 crc kubenswrapper[4744]: I1003 16:43:26.100430 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:26 crc kubenswrapper[4744]: I1003 16:43:26.961053 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-b88bn"] Oct 03 16:43:26 crc kubenswrapper[4744]: I1003 16:43:26.962728 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-b88bn" Oct 03 16:43:26 crc kubenswrapper[4744]: I1003 16:43:26.964779 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-gql6j" Oct 03 16:43:26 crc kubenswrapper[4744]: I1003 16:43:26.970703 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 03 16:43:26 crc kubenswrapper[4744]: I1003 16:43:26.971041 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 03 16:43:26 crc kubenswrapper[4744]: I1003 16:43:26.976873 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-b88bn"] Oct 03 16:43:27 crc kubenswrapper[4744]: I1003 16:43:27.020346 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lplm\" (UniqueName: \"kubernetes.io/projected/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-kube-api-access-7lplm\") pod \"neutron-db-sync-b88bn\" (UID: \"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7\") " pod="openstack/neutron-db-sync-b88bn" Oct 03 16:43:27 crc kubenswrapper[4744]: I1003 16:43:27.020637 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-config\") pod \"neutron-db-sync-b88bn\" (UID: \"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7\") " pod="openstack/neutron-db-sync-b88bn" Oct 03 16:43:27 crc kubenswrapper[4744]: I1003 16:43:27.020742 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-combined-ca-bundle\") pod \"neutron-db-sync-b88bn\" (UID: \"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7\") " pod="openstack/neutron-db-sync-b88bn" Oct 03 16:43:27 crc kubenswrapper[4744]: I1003 16:43:27.121556 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lplm\" (UniqueName: \"kubernetes.io/projected/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-kube-api-access-7lplm\") pod \"neutron-db-sync-b88bn\" (UID: \"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7\") " pod="openstack/neutron-db-sync-b88bn" Oct 03 16:43:27 crc kubenswrapper[4744]: I1003 16:43:27.121606 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-config\") pod \"neutron-db-sync-b88bn\" (UID: \"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7\") " pod="openstack/neutron-db-sync-b88bn" Oct 03 16:43:27 crc kubenswrapper[4744]: I1003 16:43:27.121632 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-combined-ca-bundle\") pod \"neutron-db-sync-b88bn\" (UID: \"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7\") " pod="openstack/neutron-db-sync-b88bn" Oct 03 16:43:27 crc kubenswrapper[4744]: I1003 16:43:27.130524 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-config\") pod \"neutron-db-sync-b88bn\" (UID: \"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7\") " pod="openstack/neutron-db-sync-b88bn" Oct 03 16:43:27 crc kubenswrapper[4744]: I1003 16:43:27.131354 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-combined-ca-bundle\") pod \"neutron-db-sync-b88bn\" (UID: \"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7\") " pod="openstack/neutron-db-sync-b88bn" Oct 03 16:43:27 crc kubenswrapper[4744]: I1003 16:43:27.140813 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lplm\" (UniqueName: \"kubernetes.io/projected/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-kube-api-access-7lplm\") pod \"neutron-db-sync-b88bn\" (UID: \"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7\") " pod="openstack/neutron-db-sync-b88bn" Oct 03 16:43:27 crc kubenswrapper[4744]: I1003 16:43:27.303508 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-b88bn" Oct 03 16:43:28 crc kubenswrapper[4744]: I1003 16:43:28.152472 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0d09-account-create-g4bpt" Oct 03 16:43:28 crc kubenswrapper[4744]: I1003 16:43:28.241341 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqfmt\" (UniqueName: \"kubernetes.io/projected/aa76346f-0ca5-4bad-8fb8-802077202e39-kube-api-access-rqfmt\") pod \"aa76346f-0ca5-4bad-8fb8-802077202e39\" (UID: \"aa76346f-0ca5-4bad-8fb8-802077202e39\") " Oct 03 16:43:28 crc kubenswrapper[4744]: I1003 16:43:28.248920 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa76346f-0ca5-4bad-8fb8-802077202e39-kube-api-access-rqfmt" (OuterVolumeSpecName: "kube-api-access-rqfmt") pod "aa76346f-0ca5-4bad-8fb8-802077202e39" (UID: "aa76346f-0ca5-4bad-8fb8-802077202e39"). InnerVolumeSpecName "kube-api-access-rqfmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:28 crc kubenswrapper[4744]: I1003 16:43:28.343846 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqfmt\" (UniqueName: \"kubernetes.io/projected/aa76346f-0ca5-4bad-8fb8-802077202e39-kube-api-access-rqfmt\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:28 crc kubenswrapper[4744]: I1003 16:43:28.719937 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0d09-account-create-g4bpt" event={"ID":"aa76346f-0ca5-4bad-8fb8-802077202e39","Type":"ContainerDied","Data":"a8fafd8891a38659d697bf71c1752465b261832828bb820f414d322d0c9b5fb4"} Oct 03 16:43:28 crc kubenswrapper[4744]: I1003 16:43:28.719975 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0d09-account-create-g4bpt" Oct 03 16:43:28 crc kubenswrapper[4744]: I1003 16:43:28.719974 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8fafd8891a38659d697bf71c1752465b261832828bb820f414d322d0c9b5fb4" Oct 03 16:43:30 crc kubenswrapper[4744]: E1003 16:43:30.051760 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Oct 03 16:43:30 crc kubenswrapper[4744]: E1003 16:43:30.052232 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5z6qc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-qbz46_openstack(7acfc27b-28b7-422c-a933-8dd2a4a99c68): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 16:43:30 crc kubenswrapper[4744]: E1003 16:43:30.053467 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-qbz46" podUID="7acfc27b-28b7-422c-a933-8dd2a4a99c68" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.186120 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3d6b-account-create-klnll" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.195727 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ffe1-account-create-nkjkf" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.198605 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.282845 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p54np\" (UniqueName: \"kubernetes.io/projected/4eeda15e-12ae-4c7c-ab47-669638562621-kube-api-access-p54np\") pod \"4eeda15e-12ae-4c7c-ab47-669638562621\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.283001 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9pn5\" (UniqueName: \"kubernetes.io/projected/c4594c72-66b3-410f-b2ef-dbb205807338-kube-api-access-v9pn5\") pod \"c4594c72-66b3-410f-b2ef-dbb205807338\" (UID: \"c4594c72-66b3-410f-b2ef-dbb205807338\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.283059 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-combined-ca-bundle\") pod \"4eeda15e-12ae-4c7c-ab47-669638562621\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.283100 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-fernet-keys\") pod \"4eeda15e-12ae-4c7c-ab47-669638562621\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.283152 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pp92\" (UniqueName: \"kubernetes.io/projected/9be7c905-97b5-40a2-b416-ff8894216fb4-kube-api-access-4pp92\") pod \"9be7c905-97b5-40a2-b416-ff8894216fb4\" (UID: \"9be7c905-97b5-40a2-b416-ff8894216fb4\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.283177 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-scripts\") pod \"4eeda15e-12ae-4c7c-ab47-669638562621\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.283220 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-config-data\") pod \"4eeda15e-12ae-4c7c-ab47-669638562621\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.283246 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-credential-keys\") pod \"4eeda15e-12ae-4c7c-ab47-669638562621\" (UID: \"4eeda15e-12ae-4c7c-ab47-669638562621\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.287638 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eeda15e-12ae-4c7c-ab47-669638562621-kube-api-access-p54np" (OuterVolumeSpecName: "kube-api-access-p54np") pod "4eeda15e-12ae-4c7c-ab47-669638562621" (UID: "4eeda15e-12ae-4c7c-ab47-669638562621"). InnerVolumeSpecName "kube-api-access-p54np". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.288507 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4594c72-66b3-410f-b2ef-dbb205807338-kube-api-access-v9pn5" (OuterVolumeSpecName: "kube-api-access-v9pn5") pod "c4594c72-66b3-410f-b2ef-dbb205807338" (UID: "c4594c72-66b3-410f-b2ef-dbb205807338"). InnerVolumeSpecName "kube-api-access-v9pn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.288989 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4eeda15e-12ae-4c7c-ab47-669638562621" (UID: "4eeda15e-12ae-4c7c-ab47-669638562621"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.291060 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-scripts" (OuterVolumeSpecName: "scripts") pod "4eeda15e-12ae-4c7c-ab47-669638562621" (UID: "4eeda15e-12ae-4c7c-ab47-669638562621"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.294065 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4eeda15e-12ae-4c7c-ab47-669638562621" (UID: "4eeda15e-12ae-4c7c-ab47-669638562621"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.297665 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9be7c905-97b5-40a2-b416-ff8894216fb4-kube-api-access-4pp92" (OuterVolumeSpecName: "kube-api-access-4pp92") pod "9be7c905-97b5-40a2-b416-ff8894216fb4" (UID: "9be7c905-97b5-40a2-b416-ff8894216fb4"). InnerVolumeSpecName "kube-api-access-4pp92". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.315629 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4eeda15e-12ae-4c7c-ab47-669638562621" (UID: "4eeda15e-12ae-4c7c-ab47-669638562621"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.335829 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-config-data" (OuterVolumeSpecName: "config-data") pod "4eeda15e-12ae-4c7c-ab47-669638562621" (UID: "4eeda15e-12ae-4c7c-ab47-669638562621"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.387141 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9pn5\" (UniqueName: \"kubernetes.io/projected/c4594c72-66b3-410f-b2ef-dbb205807338-kube-api-access-v9pn5\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.387179 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.387194 4744 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.387205 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pp92\" (UniqueName: \"kubernetes.io/projected/9be7c905-97b5-40a2-b416-ff8894216fb4-kube-api-access-4pp92\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.387216 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.387226 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.387236 4744 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4eeda15e-12ae-4c7c-ab47-669638562621-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.387244 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p54np\" (UniqueName: \"kubernetes.io/projected/4eeda15e-12ae-4c7c-ab47-669638562621-kube-api-access-p54np\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.389176 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.487795 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2thc\" (UniqueName: \"kubernetes.io/projected/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-kube-api-access-c2thc\") pod \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.488086 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-combined-ca-bundle\") pod \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.488134 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.488183 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-logs\") pod \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.488208 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-ceph\") pod \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.488247 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-config-data\") pod \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.488324 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-scripts\") pod \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.488408 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-httpd-run\") pod \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\" (UID: \"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.489094 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" (UID: "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.492721 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-logs" (OuterVolumeSpecName: "logs") pod "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" (UID: "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.495670 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-scripts" (OuterVolumeSpecName: "scripts") pod "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" (UID: "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.496688 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-kube-api-access-c2thc" (OuterVolumeSpecName: "kube-api-access-c2thc") pod "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" (UID: "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc"). InnerVolumeSpecName "kube-api-access-c2thc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.498742 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" (UID: "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.502874 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-ceph" (OuterVolumeSpecName: "ceph") pod "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" (UID: "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.557848 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.590833 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2thc\" (UniqueName: \"kubernetes.io/projected/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-kube-api-access-c2thc\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.590880 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.590895 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.590906 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.590916 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.590926 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.593654 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" (UID: "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.623246 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.626504 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-config-data" (OuterVolumeSpecName: "config-data") pod "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" (UID: "8ea987f3-4a2b-4f52-a5ef-104dc793f2fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.646417 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.692443 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-httpd-run\") pod \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.692525 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-dns-swift-storage-0\") pod \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.692549 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q52cx\" (UniqueName: \"kubernetes.io/projected/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-kube-api-access-q52cx\") pod \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.692582 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-combined-ca-bundle\") pod \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.692621 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-logs\") pod \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.692680 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-config-data\") pod \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.692709 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-config\") pod \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.692724 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-ceph\") pod \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.692752 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w95p9\" (UniqueName: \"kubernetes.io/projected/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-kube-api-access-w95p9\") pod \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.692774 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-scripts\") pod \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.692802 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-ovsdbserver-sb\") pod \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.692834 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\" (UID: \"1d2b85de-e61a-4a88-8180-c0cc398b7e6d\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.692857 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-ovsdbserver-nb\") pod \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.692907 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-dns-svc\") pod \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\" (UID: \"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914\") " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.692933 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1d2b85de-e61a-4a88-8180-c0cc398b7e6d" (UID: "1d2b85de-e61a-4a88-8180-c0cc398b7e6d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.693131 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-logs" (OuterVolumeSpecName: "logs") pod "1d2b85de-e61a-4a88-8180-c0cc398b7e6d" (UID: "1d2b85de-e61a-4a88-8180-c0cc398b7e6d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.693260 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.693271 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.693280 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.693288 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.693296 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.697439 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-scripts" (OuterVolumeSpecName: "scripts") pod "1d2b85de-e61a-4a88-8180-c0cc398b7e6d" (UID: "1d2b85de-e61a-4a88-8180-c0cc398b7e6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.698571 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-ceph" (OuterVolumeSpecName: "ceph") pod "1d2b85de-e61a-4a88-8180-c0cc398b7e6d" (UID: "1d2b85de-e61a-4a88-8180-c0cc398b7e6d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.699149 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "1d2b85de-e61a-4a88-8180-c0cc398b7e6d" (UID: "1d2b85de-e61a-4a88-8180-c0cc398b7e6d"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.702814 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-kube-api-access-w95p9" (OuterVolumeSpecName: "kube-api-access-w95p9") pod "1d2b85de-e61a-4a88-8180-c0cc398b7e6d" (UID: "1d2b85de-e61a-4a88-8180-c0cc398b7e6d"). InnerVolumeSpecName "kube-api-access-w95p9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.702955 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-kube-api-access-q52cx" (OuterVolumeSpecName: "kube-api-access-q52cx") pod "3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" (UID: "3c4651bb-32f8-4baa-b0a3-a8fd37ca9914"). InnerVolumeSpecName "kube-api-access-q52cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.738679 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dc6d57fbc-4s82v" event={"ID":"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc","Type":"ContainerStarted","Data":"49414963259aa5d89aa752ae2d0e34de99d22af2fd18f7e139dd5e7d6c1886af"} Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.743116 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1d2b85de-e61a-4a88-8180-c0cc398b7e6d","Type":"ContainerDied","Data":"3df562a2b97659cf0c6868538a7fe8b42b6f16622344768ac77ea0a67e184c97"} Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.743181 4744 scope.go:117] "RemoveContainer" containerID="2ecdf210efa873bc6811289e1fba3768d387757a380c9253b160612617cd70ef" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.743352 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.744396 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d2b85de-e61a-4a88-8180-c0cc398b7e6d" (UID: "1d2b85de-e61a-4a88-8180-c0cc398b7e6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.747776 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4","Type":"ContainerStarted","Data":"c4f9d07a32023e6cae16d77eac1219e6d66fcd4e519e8880aed981dd3e712ebe"} Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.749587 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3d6b-account-create-klnll" event={"ID":"9be7c905-97b5-40a2-b416-ff8894216fb4","Type":"ContainerDied","Data":"df72c96120b22a6825e817ce75f8450897fc4f9fbda38e35d5fbfb7224a5544a"} Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.749613 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df72c96120b22a6825e817ce75f8450897fc4f9fbda38e35d5fbfb7224a5544a" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.749683 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3d6b-account-create-klnll" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.753138 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-ffe1-account-create-nkjkf" event={"ID":"c4594c72-66b3-410f-b2ef-dbb205807338","Type":"ContainerDied","Data":"9d642b49d4045879438badbedb672b316c8255f4b024649ce76b3aecb889e76d"} Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.753180 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d642b49d4045879438badbedb672b316c8255f4b024649ce76b3aecb889e76d" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.753236 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ffe1-account-create-nkjkf" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.755020 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-csk6k" event={"ID":"4eeda15e-12ae-4c7c-ab47-669638562621","Type":"ContainerDied","Data":"858cbc1446284799bf597619ad46aec968c53311c5a9196db303cd6049b90686"} Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.755058 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="858cbc1446284799bf597619ad46aec968c53311c5a9196db303cd6049b90686" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.755112 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-csk6k" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.766756 4744 scope.go:117] "RemoveContainer" containerID="a05cc65f56c3d394943b07d758b31ef01f67118f473ee2b9c3bb51af31214e3e" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.769249 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79c99f65c-jjfnc" event={"ID":"ecb56805-c2cc-4736-b5bc-84440acb7fa6","Type":"ContainerStarted","Data":"dc4393a029969f5d1dded83e80cefa58f4efd64de2309c8dab88639cb4b8bce6"} Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.785801 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" event={"ID":"3c4651bb-32f8-4baa-b0a3-a8fd37ca9914","Type":"ContainerDied","Data":"505cedd02078765974fb0cf966b86f4451231120d88a429acc9c92cbd2a8dd38"} Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.785890 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-8hrkj" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.791425 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.791576 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8ea987f3-4a2b-4f52-a5ef-104dc793f2fc","Type":"ContainerDied","Data":"be311d43401ab3a383990ec423ff8c5ea94f49668d7cd94fab5fcd57937b3af6"} Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.794619 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.794651 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w95p9\" (UniqueName: \"kubernetes.io/projected/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-kube-api-access-w95p9\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.794663 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.794700 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.794715 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q52cx\" (UniqueName: \"kubernetes.io/projected/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-kube-api-access-q52cx\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.794727 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.799809 4744 scope.go:117] "RemoveContainer" containerID="1228505882f8ef70a3feeb4a8aecc382e7742259e6fb684543f703f4163038a5" Oct 03 16:43:30 crc kubenswrapper[4744]: E1003 16:43:30.800079 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-qbz46" podUID="7acfc27b-28b7-422c-a933-8dd2a4a99c68" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.803451 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" (UID: "3c4651bb-32f8-4baa-b0a3-a8fd37ca9914"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.826618 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-config-data" (OuterVolumeSpecName: "config-data") pod "1d2b85de-e61a-4a88-8180-c0cc398b7e6d" (UID: "1d2b85de-e61a-4a88-8180-c0cc398b7e6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.828477 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.829216 4744 scope.go:117] "RemoveContainer" containerID="e953d762fe5d2bd3ded9de739a090ecbb8970d7fe582e876cf8a5fc71aea20f1" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.831160 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" (UID: "3c4651bb-32f8-4baa-b0a3-a8fd37ca9914"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.838167 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" (UID: "3c4651bb-32f8-4baa-b0a3-a8fd37ca9914"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.838915 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-config" (OuterVolumeSpecName: "config") pod "3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" (UID: "3c4651bb-32f8-4baa-b0a3-a8fd37ca9914"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.848188 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" (UID: "3c4651bb-32f8-4baa-b0a3-a8fd37ca9914"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.854562 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.862976 4744 scope.go:117] "RemoveContainer" containerID="1023f99e320002af835fecc4de9f9781c1fd5de2d6dbba843abdf1fe761a501e" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.881189 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.890381 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 16:43:30 crc kubenswrapper[4744]: E1003 16:43:30.890925 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d2b85de-e61a-4a88-8180-c0cc398b7e6d" containerName="glance-httpd" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.890939 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d2b85de-e61a-4a88-8180-c0cc398b7e6d" containerName="glance-httpd" Oct 03 16:43:30 crc kubenswrapper[4744]: E1003 16:43:30.890967 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" containerName="init" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.890973 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" containerName="init" Oct 03 16:43:30 crc kubenswrapper[4744]: E1003 16:43:30.890988 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" containerName="glance-log" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.890994 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" containerName="glance-log" Oct 03 16:43:30 crc kubenswrapper[4744]: E1003 16:43:30.891004 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eeda15e-12ae-4c7c-ab47-669638562621" containerName="keystone-bootstrap" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891011 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eeda15e-12ae-4c7c-ab47-669638562621" containerName="keystone-bootstrap" Oct 03 16:43:30 crc kubenswrapper[4744]: E1003 16:43:30.891040 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9be7c905-97b5-40a2-b416-ff8894216fb4" containerName="mariadb-account-create" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891048 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9be7c905-97b5-40a2-b416-ff8894216fb4" containerName="mariadb-account-create" Oct 03 16:43:30 crc kubenswrapper[4744]: E1003 16:43:30.891063 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4594c72-66b3-410f-b2ef-dbb205807338" containerName="mariadb-account-create" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891071 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4594c72-66b3-410f-b2ef-dbb205807338" containerName="mariadb-account-create" Oct 03 16:43:30 crc kubenswrapper[4744]: E1003 16:43:30.891089 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa76346f-0ca5-4bad-8fb8-802077202e39" containerName="mariadb-account-create" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891095 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa76346f-0ca5-4bad-8fb8-802077202e39" containerName="mariadb-account-create" Oct 03 16:43:30 crc kubenswrapper[4744]: E1003 16:43:30.891126 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" containerName="dnsmasq-dns" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891132 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" containerName="dnsmasq-dns" Oct 03 16:43:30 crc kubenswrapper[4744]: E1003 16:43:30.891143 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d2b85de-e61a-4a88-8180-c0cc398b7e6d" containerName="glance-log" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891150 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d2b85de-e61a-4a88-8180-c0cc398b7e6d" containerName="glance-log" Oct 03 16:43:30 crc kubenswrapper[4744]: E1003 16:43:30.891165 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" containerName="glance-httpd" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891170 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" containerName="glance-httpd" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891391 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa76346f-0ca5-4bad-8fb8-802077202e39" containerName="mariadb-account-create" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891413 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9be7c905-97b5-40a2-b416-ff8894216fb4" containerName="mariadb-account-create" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891453 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" containerName="dnsmasq-dns" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891465 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eeda15e-12ae-4c7c-ab47-669638562621" containerName="keystone-bootstrap" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891486 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4594c72-66b3-410f-b2ef-dbb205807338" containerName="mariadb-account-create" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891534 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d2b85de-e61a-4a88-8180-c0cc398b7e6d" containerName="glance-httpd" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891557 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d2b85de-e61a-4a88-8180-c0cc398b7e6d" containerName="glance-log" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891572 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" containerName="glance-log" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.891583 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" containerName="glance-httpd" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.903885 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.906247 4744 scope.go:117] "RemoveContainer" containerID="ab06515926dcd29c8cf7aebb1c3a15cfa8fc923a716d85ca72850c8e24339a23" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.918322 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.919143 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.958599 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ea987f3-4a2b-4f52-a5ef-104dc793f2fc" path="/var/lib/kubelet/pods/8ea987f3-4a2b-4f52-a5ef-104dc793f2fc/volumes" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.959735 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-b88bn"] Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.959760 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85d8c4f7b-9pn28"] Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.959785 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.960790 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d2b85de-e61a-4a88-8180-c0cc398b7e6d-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.960824 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.960839 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.960857 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.960867 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.960877 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.960887 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:30 crc kubenswrapper[4744]: I1003 16:43:30.992702 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7978488848-g4fcf"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.002142 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-75bcc4d5c-s79cg"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.063172 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.063266 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwdpp\" (UniqueName: \"kubernetes.io/projected/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-kube-api-access-xwdpp\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.063365 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-ceph\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.063395 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.063416 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-scripts\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.063431 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.063447 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-logs\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.063467 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-config-data\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.063487 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.164857 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.164925 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwdpp\" (UniqueName: \"kubernetes.io/projected/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-kube-api-access-xwdpp\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.164982 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-ceph\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.165023 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.165047 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-scripts\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.165060 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.165074 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-logs\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.165094 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-config-data\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.165111 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.165649 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.167386 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.167735 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-logs\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.169182 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.169680 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-ceph\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.170097 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.173374 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-config-data\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.174144 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-scripts\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.191379 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwdpp\" (UniqueName: \"kubernetes.io/projected/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-kube-api-access-xwdpp\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.201744 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.230987 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.406044 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-csk6k"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.425760 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-csk6k"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.434423 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-8hrkj"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.450820 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-8hrkj"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.478764 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.490271 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.500785 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.502769 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.507748 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.507925 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.510467 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.550944 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-9ftfh"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.552294 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.557966 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.558335 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2z2zg" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.558581 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.558758 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.597890 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9ftfh"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.647983 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-6k6b8"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.649344 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6k6b8" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.659343 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.665806 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-ttscr" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.679572 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.679842 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.679920 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-scripts\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.679992 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb8803f7-c116-4175-b653-4b451e1da26d-logs\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.680086 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.680218 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-config-data\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.680286 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cb8803f7-c116-4175-b653-4b451e1da26d-ceph\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.680383 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.680462 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsgnv\" (UniqueName: \"kubernetes.io/projected/cb8803f7-c116-4175-b653-4b451e1da26d-kube-api-access-lsgnv\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.680698 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wqvw\" (UniqueName: \"kubernetes.io/projected/245dc938-9219-43e5-8e69-10e033862451-kube-api-access-2wqvw\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.680766 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-combined-ca-bundle\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.680844 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb8803f7-c116-4175-b653-4b451e1da26d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.681061 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-fernet-keys\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.681144 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-credential-keys\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.681233 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.684688 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-6k6b8"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782516 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782552 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782581 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-scripts\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782603 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb8803f7-c116-4175-b653-4b451e1da26d-logs\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782618 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782637 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqlt9\" (UniqueName: \"kubernetes.io/projected/dc8def8d-6149-4797-b453-103ec15579f6-kube-api-access-wqlt9\") pod \"barbican-db-sync-6k6b8\" (UID: \"dc8def8d-6149-4797-b453-103ec15579f6\") " pod="openstack/barbican-db-sync-6k6b8" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782664 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-config-data\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782680 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cb8803f7-c116-4175-b653-4b451e1da26d-ceph\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782700 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782721 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsgnv\" (UniqueName: \"kubernetes.io/projected/cb8803f7-c116-4175-b653-4b451e1da26d-kube-api-access-lsgnv\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782749 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc8def8d-6149-4797-b453-103ec15579f6-combined-ca-bundle\") pod \"barbican-db-sync-6k6b8\" (UID: \"dc8def8d-6149-4797-b453-103ec15579f6\") " pod="openstack/barbican-db-sync-6k6b8" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782771 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wqvw\" (UniqueName: \"kubernetes.io/projected/245dc938-9219-43e5-8e69-10e033862451-kube-api-access-2wqvw\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782790 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-combined-ca-bundle\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782817 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb8803f7-c116-4175-b653-4b451e1da26d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782852 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dc8def8d-6149-4797-b453-103ec15579f6-db-sync-config-data\") pod \"barbican-db-sync-6k6b8\" (UID: \"dc8def8d-6149-4797-b453-103ec15579f6\") " pod="openstack/barbican-db-sync-6k6b8" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782873 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-fernet-keys\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782893 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-credential-keys\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.782932 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.784088 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.784485 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb8803f7-c116-4175-b653-4b451e1da26d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.785267 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb8803f7-c116-4175-b653-4b451e1da26d-logs\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.792034 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-credential-keys\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.794997 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-combined-ca-bundle\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.803852 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-scripts\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.805426 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-fernet-keys\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.811311 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.811369 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-8bj8c"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.812647 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-8bj8c" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.814039 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.816595 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.821094 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-h7rcp" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.821910 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.826853 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cb8803f7-c116-4175-b653-4b451e1da26d-ceph\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.827062 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-krczc"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.826852 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-config-data\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.827311 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.829378 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.835113 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.835332 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8r7z7" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.837309 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-8bj8c"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.838209 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wqvw\" (UniqueName: \"kubernetes.io/projected/245dc938-9219-43e5-8e69-10e033862451-kube-api-access-2wqvw\") pod \"keystone-bootstrap-9ftfh\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.840979 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsgnv\" (UniqueName: \"kubernetes.io/projected/cb8803f7-c116-4175-b653-4b451e1da26d-kube-api-access-lsgnv\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.845346 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.856037 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-krczc"] Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.866114 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.869027 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7978488848-g4fcf" event={"ID":"874834f7-8c7c-422d-9ec1-822d591f6d75","Type":"ContainerStarted","Data":"6ea220f0d199dc62a2062ec5f75b92116abceb965c5ab92e91a2b31c6d396809"} Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.869084 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7978488848-g4fcf" event={"ID":"874834f7-8c7c-422d-9ec1-822d591f6d75","Type":"ContainerStarted","Data":"5ceea66d6ac67e106ac5564e62972e6abb6747edf429191f7261cb8fbec2fecb"} Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.874556 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85d8c4f7b-9pn28" event={"ID":"39c711a0-3cea-4091-94a0-ec9df6e6a2d8","Type":"ContainerStarted","Data":"9725601278de2c074e46f2b47d3a9100b64e3097a5b012a9add35c50995ea1d8"} Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.874612 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85d8c4f7b-9pn28" event={"ID":"39c711a0-3cea-4091-94a0-ec9df6e6a2d8","Type":"ContainerStarted","Data":"0e4679acef225bd70082b9ebe2f82562c8ae6454e4cf3f90aa333b6c2d464223"} Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.897242 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dc6d57fbc-4s82v" event={"ID":"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc","Type":"ContainerStarted","Data":"25891acef34e96920974d576db823a20cd67c230f770dbc5f5e95ef8ee95fc40"} Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.897436 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-dc6d57fbc-4s82v" podUID="bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" containerName="horizon-log" containerID="cri-o://49414963259aa5d89aa752ae2d0e34de99d22af2fd18f7e139dd5e7d6c1886af" gracePeriod=30 Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.898615 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-dc6d57fbc-4s82v" podUID="bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" containerName="horizon" containerID="cri-o://25891acef34e96920974d576db823a20cd67c230f770dbc5f5e95ef8ee95fc40" gracePeriod=30 Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.911833 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-b88bn" event={"ID":"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7","Type":"ContainerStarted","Data":"6286e9ec39d254af81ef9ab41125f9cdbd2ce065a205572c2bbe1eb4f3a4bd0f"} Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.912098 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-b88bn" event={"ID":"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7","Type":"ContainerStarted","Data":"c7ab859003cda6ba15c877d5259b1371534c24a3c3911044303c6a061b1ec6c4"} Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.915842 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-combined-ca-bundle\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.915890 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99bm5\" (UniqueName: \"kubernetes.io/projected/a8054599-30d8-42e0-bf45-2a8758b10486-kube-api-access-99bm5\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.915916 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dc8def8d-6149-4797-b453-103ec15579f6-db-sync-config-data\") pod \"barbican-db-sync-6k6b8\" (UID: \"dc8def8d-6149-4797-b453-103ec15579f6\") " pod="openstack/barbican-db-sync-6k6b8" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.916028 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-config-data\") pod \"manila-db-sync-8bj8c\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " pod="openstack/manila-db-sync-8bj8c" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.916045 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-scripts\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.916123 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58hb9\" (UniqueName: \"kubernetes.io/projected/df7dff0d-99e0-4e98-b752-9096d24e314b-kube-api-access-58hb9\") pod \"manila-db-sync-8bj8c\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " pod="openstack/manila-db-sync-8bj8c" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.916232 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-config-data\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.916255 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqlt9\" (UniqueName: \"kubernetes.io/projected/dc8def8d-6149-4797-b453-103ec15579f6-kube-api-access-wqlt9\") pod \"barbican-db-sync-6k6b8\" (UID: \"dc8def8d-6149-4797-b453-103ec15579f6\") " pod="openstack/barbican-db-sync-6k6b8" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.916320 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8054599-30d8-42e0-bf45-2a8758b10486-etc-machine-id\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.916373 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc8def8d-6149-4797-b453-103ec15579f6-combined-ca-bundle\") pod \"barbican-db-sync-6k6b8\" (UID: \"dc8def8d-6149-4797-b453-103ec15579f6\") " pod="openstack/barbican-db-sync-6k6b8" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.916399 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-db-sync-config-data\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.916435 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-job-config-data\") pod \"manila-db-sync-8bj8c\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " pod="openstack/manila-db-sync-8bj8c" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.916466 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-combined-ca-bundle\") pod \"manila-db-sync-8bj8c\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " pod="openstack/manila-db-sync-8bj8c" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.918338 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.924937 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75bcc4d5c-s79cg" event={"ID":"d991f026-62e5-4515-924d-ae7267cab11f","Type":"ContainerStarted","Data":"e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7"} Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.924978 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75bcc4d5c-s79cg" event={"ID":"d991f026-62e5-4515-924d-ae7267cab11f","Type":"ContainerStarted","Data":"61f5c36bba91ff88181f6563fedd5e86a33d4a733ff0dd6a9b10ebfd4202c48c"} Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.926577 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dc8def8d-6149-4797-b453-103ec15579f6-db-sync-config-data\") pod \"barbican-db-sync-6k6b8\" (UID: \"dc8def8d-6149-4797-b453-103ec15579f6\") " pod="openstack/barbican-db-sync-6k6b8" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.928216 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79c99f65c-jjfnc" event={"ID":"ecb56805-c2cc-4736-b5bc-84440acb7fa6","Type":"ContainerStarted","Data":"d9e099f2d91d174ed4080ff290be3a0e907e1d6ce44212fba5bdea7f9faf6350"} Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.932403 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-dc6d57fbc-4s82v" podStartSLOduration=3.141724016 podStartE2EDuration="17.932387444s" podCreationTimestamp="2025-10-03 16:43:14 +0000 UTC" firstStartedPulling="2025-10-03 16:43:15.497142965 +0000 UTC m=+1121.777018861" lastFinishedPulling="2025-10-03 16:43:30.287806393 +0000 UTC m=+1136.567682289" observedRunningTime="2025-10-03 16:43:31.922587662 +0000 UTC m=+1138.202463558" watchObservedRunningTime="2025-10-03 16:43:31.932387444 +0000 UTC m=+1138.212263330" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.944357 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqlt9\" (UniqueName: \"kubernetes.io/projected/dc8def8d-6149-4797-b453-103ec15579f6-kube-api-access-wqlt9\") pod \"barbican-db-sync-6k6b8\" (UID: \"dc8def8d-6149-4797-b453-103ec15579f6\") " pod="openstack/barbican-db-sync-6k6b8" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.944675 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79c99f65c-jjfnc" podUID="ecb56805-c2cc-4736-b5bc-84440acb7fa6" containerName="horizon-log" containerID="cri-o://dc4393a029969f5d1dded83e80cefa58f4efd64de2309c8dab88639cb4b8bce6" gracePeriod=30 Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.946972 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-79c99f65c-jjfnc" podUID="ecb56805-c2cc-4736-b5bc-84440acb7fa6" containerName="horizon" containerID="cri-o://d9e099f2d91d174ed4080ff290be3a0e907e1d6ce44212fba5bdea7f9faf6350" gracePeriod=30 Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.963286 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc8def8d-6149-4797-b453-103ec15579f6-combined-ca-bundle\") pod \"barbican-db-sync-6k6b8\" (UID: \"dc8def8d-6149-4797-b453-103ec15579f6\") " pod="openstack/barbican-db-sync-6k6b8" Oct 03 16:43:31 crc kubenswrapper[4744]: I1003 16:43:31.978576 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-b88bn" podStartSLOduration=5.9785547 podStartE2EDuration="5.9785547s" podCreationTimestamp="2025-10-03 16:43:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:31.952603802 +0000 UTC m=+1138.232479718" watchObservedRunningTime="2025-10-03 16:43:31.9785547 +0000 UTC m=+1138.258430596" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:31.998991 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-79c99f65c-jjfnc" podStartSLOduration=3.900235571 podStartE2EDuration="18.998973943s" podCreationTimestamp="2025-10-03 16:43:13 +0000 UTC" firstStartedPulling="2025-10-03 16:43:15.150609689 +0000 UTC m=+1121.430485585" lastFinishedPulling="2025-10-03 16:43:30.249348061 +0000 UTC m=+1136.529223957" observedRunningTime="2025-10-03 16:43:31.981059587 +0000 UTC m=+1138.260935483" watchObservedRunningTime="2025-10-03 16:43:31.998973943 +0000 UTC m=+1138.278849839" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.021244 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-config-data\") pod \"manila-db-sync-8bj8c\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " pod="openstack/manila-db-sync-8bj8c" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.021285 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-scripts\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.021342 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58hb9\" (UniqueName: \"kubernetes.io/projected/df7dff0d-99e0-4e98-b752-9096d24e314b-kube-api-access-58hb9\") pod \"manila-db-sync-8bj8c\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " pod="openstack/manila-db-sync-8bj8c" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.021390 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-config-data\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.021476 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8054599-30d8-42e0-bf45-2a8758b10486-etc-machine-id\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.021544 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-db-sync-config-data\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.021576 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-job-config-data\") pod \"manila-db-sync-8bj8c\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " pod="openstack/manila-db-sync-8bj8c" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.021605 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-combined-ca-bundle\") pod \"manila-db-sync-8bj8c\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " pod="openstack/manila-db-sync-8bj8c" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.021696 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-combined-ca-bundle\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.021713 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99bm5\" (UniqueName: \"kubernetes.io/projected/a8054599-30d8-42e0-bf45-2a8758b10486-kube-api-access-99bm5\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.025412 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8054599-30d8-42e0-bf45-2a8758b10486-etc-machine-id\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.030382 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-db-sync-config-data\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.031242 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-config-data\") pod \"manila-db-sync-8bj8c\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " pod="openstack/manila-db-sync-8bj8c" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.031507 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-job-config-data\") pod \"manila-db-sync-8bj8c\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " pod="openstack/manila-db-sync-8bj8c" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.032160 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-combined-ca-bundle\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.034238 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-combined-ca-bundle\") pod \"manila-db-sync-8bj8c\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " pod="openstack/manila-db-sync-8bj8c" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.046214 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99bm5\" (UniqueName: \"kubernetes.io/projected/a8054599-30d8-42e0-bf45-2a8758b10486-kube-api-access-99bm5\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.047734 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-config-data\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.057036 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58hb9\" (UniqueName: \"kubernetes.io/projected/df7dff0d-99e0-4e98-b752-9096d24e314b-kube-api-access-58hb9\") pod \"manila-db-sync-8bj8c\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " pod="openstack/manila-db-sync-8bj8c" Oct 03 16:43:32 crc kubenswrapper[4744]: W1003 16:43:32.078682 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf524b0a3_3650_4f2e_9a27_d74b82e3e06d.slice/crio-31a435ffe66addb6d519f50e0249a245a72d47ac07255246f9eb9501ff114a43 WatchSource:0}: Error finding container 31a435ffe66addb6d519f50e0249a245a72d47ac07255246f9eb9501ff114a43: Status 404 returned error can't find the container with id 31a435ffe66addb6d519f50e0249a245a72d47ac07255246f9eb9501ff114a43 Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.091537 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-scripts\") pod \"cinder-db-sync-krczc\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.094736 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.116890 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6k6b8" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.158173 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.167137 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-8bj8c" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.197595 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-krczc" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.569984 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9ftfh"] Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.904323 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d2b85de-e61a-4a88-8180-c0cc398b7e6d" path="/var/lib/kubelet/pods/1d2b85de-e61a-4a88-8180-c0cc398b7e6d/volumes" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.905159 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c4651bb-32f8-4baa-b0a3-a8fd37ca9914" path="/var/lib/kubelet/pods/3c4651bb-32f8-4baa-b0a3-a8fd37ca9914/volumes" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.973988 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-75bcc4d5c-s79cg" podUID="d991f026-62e5-4515-924d-ae7267cab11f" containerName="horizon-log" containerID="cri-o://e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7" gracePeriod=30 Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.974264 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-75bcc4d5c-s79cg" podUID="d991f026-62e5-4515-924d-ae7267cab11f" containerName="horizon" containerID="cri-o://1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3" gracePeriod=30 Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.989460 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eeda15e-12ae-4c7c-ab47-669638562621" path="/var/lib/kubelet/pods/4eeda15e-12ae-4c7c-ab47-669638562621/volumes" Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.990085 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7978488848-g4fcf" event={"ID":"874834f7-8c7c-422d-9ec1-822d591f6d75","Type":"ContainerStarted","Data":"3fd2733e66910ca5adf8f373718198684ccaa21a7e01b73dbd60efbc2cae7e6c"} Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.990121 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85d8c4f7b-9pn28" event={"ID":"39c711a0-3cea-4091-94a0-ec9df6e6a2d8","Type":"ContainerStarted","Data":"908ee237dace1930ccf6b4f750ca12b8461daace0aa87e0b5312b59d86e1937d"} Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.990152 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75bcc4d5c-s79cg" event={"ID":"d991f026-62e5-4515-924d-ae7267cab11f","Type":"ContainerStarted","Data":"1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3"} Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.990165 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f524b0a3-3650-4f2e-9a27-d74b82e3e06d","Type":"ContainerStarted","Data":"31a435ffe66addb6d519f50e0249a245a72d47ac07255246f9eb9501ff114a43"} Oct 03 16:43:32 crc kubenswrapper[4744]: I1003 16:43:32.996921 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7978488848-g4fcf" podStartSLOduration=7.996904305 podStartE2EDuration="7.996904305s" podCreationTimestamp="2025-10-03 16:43:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:32.992524106 +0000 UTC m=+1139.272400012" watchObservedRunningTime="2025-10-03 16:43:32.996904305 +0000 UTC m=+1139.276780201" Oct 03 16:43:33 crc kubenswrapper[4744]: I1003 16:43:33.029738 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-75bcc4d5c-s79cg" podStartSLOduration=15.029720359 podStartE2EDuration="15.029720359s" podCreationTimestamp="2025-10-03 16:43:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:33.022085286 +0000 UTC m=+1139.301961182" watchObservedRunningTime="2025-10-03 16:43:33.029720359 +0000 UTC m=+1139.309596255" Oct 03 16:43:33 crc kubenswrapper[4744]: I1003 16:43:33.080677 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-85d8c4f7b-9pn28" podStartSLOduration=8.080657944 podStartE2EDuration="8.080657944s" podCreationTimestamp="2025-10-03 16:43:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:33.070410082 +0000 UTC m=+1139.350285978" watchObservedRunningTime="2025-10-03 16:43:33.080657944 +0000 UTC m=+1139.360533840" Oct 03 16:43:34 crc kubenswrapper[4744]: I1003 16:43:34.041727 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9ftfh" event={"ID":"245dc938-9219-43e5-8e69-10e033862451","Type":"ContainerStarted","Data":"dee8d427f80c0aa195cab8dc7f68e0984ec3b09e2ac8660402c5ed9b553cd581"} Oct 03 16:43:34 crc kubenswrapper[4744]: I1003 16:43:34.311149 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:43:34 crc kubenswrapper[4744]: I1003 16:43:34.578154 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 16:43:34 crc kubenswrapper[4744]: I1003 16:43:34.630579 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-krczc"] Oct 03 16:43:34 crc kubenswrapper[4744]: I1003 16:43:34.651474 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-6k6b8"] Oct 03 16:43:34 crc kubenswrapper[4744]: I1003 16:43:34.662216 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-8bj8c"] Oct 03 16:43:34 crc kubenswrapper[4744]: I1003 16:43:34.668793 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:43:34 crc kubenswrapper[4744]: I1003 16:43:34.668851 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:43:34 crc kubenswrapper[4744]: I1003 16:43:34.677660 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:43:35 crc kubenswrapper[4744]: I1003 16:43:35.095546 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9ftfh" event={"ID":"245dc938-9219-43e5-8e69-10e033862451","Type":"ContainerStarted","Data":"0802af45ac185ed4e7848dce58259daaefb0b25cf43bdc9fae87eebd75be34ed"} Oct 03 16:43:35 crc kubenswrapper[4744]: I1003 16:43:35.103241 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb8803f7-c116-4175-b653-4b451e1da26d","Type":"ContainerStarted","Data":"a392bc7af180766ab183d06d2805e964217faea000af24f9c19845fae717be91"} Oct 03 16:43:35 crc kubenswrapper[4744]: I1003 16:43:35.108341 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6k6b8" event={"ID":"dc8def8d-6149-4797-b453-103ec15579f6","Type":"ContainerStarted","Data":"13aba0b99f1560496ecb508e3b1c42067ec0c4d2dcdfe7b1e50937a5be6a2716"} Oct 03 16:43:35 crc kubenswrapper[4744]: I1003 16:43:35.109951 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-8bj8c" event={"ID":"df7dff0d-99e0-4e98-b752-9096d24e314b","Type":"ContainerStarted","Data":"6553ff43656eeed87600ec11e174aab60f5d90c23e020ab2782ca19348b05eff"} Oct 03 16:43:35 crc kubenswrapper[4744]: I1003 16:43:35.112874 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f524b0a3-3650-4f2e-9a27-d74b82e3e06d","Type":"ContainerStarted","Data":"502896ddc00c75f7a96640a4e1f09fb28ca1e152a7bd09d0d12cd4ecdf2ac7be"} Oct 03 16:43:35 crc kubenswrapper[4744]: I1003 16:43:35.114353 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-krczc" event={"ID":"a8054599-30d8-42e0-bf45-2a8758b10486","Type":"ContainerStarted","Data":"cf120b7b85db73f9b55659238209e60a464ef5438c7aa01a1a2e471e28c33d7d"} Oct 03 16:43:35 crc kubenswrapper[4744]: I1003 16:43:35.115476 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4","Type":"ContainerStarted","Data":"7eb661e5353163eb93dc19558c766a2c4d32ad30ebec5298fa08a7861f1a24d1"} Oct 03 16:43:35 crc kubenswrapper[4744]: I1003 16:43:35.157913 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-9ftfh" podStartSLOduration=4.157894762 podStartE2EDuration="4.157894762s" podCreationTimestamp="2025-10-03 16:43:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:35.135001033 +0000 UTC m=+1141.414876919" watchObservedRunningTime="2025-10-03 16:43:35.157894762 +0000 UTC m=+1141.437770648" Oct 03 16:43:36 crc kubenswrapper[4744]: I1003 16:43:36.015976 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:36 crc kubenswrapper[4744]: I1003 16:43:36.017185 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:43:36 crc kubenswrapper[4744]: I1003 16:43:36.100583 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:36 crc kubenswrapper[4744]: I1003 16:43:36.100643 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:36 crc kubenswrapper[4744]: I1003 16:43:36.147945 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f524b0a3-3650-4f2e-9a27-d74b82e3e06d","Type":"ContainerStarted","Data":"85cb0fee8dd78b21ef8c932996f2e53b5687f54c2bceb977c555d98b63cba81d"} Oct 03 16:43:36 crc kubenswrapper[4744]: I1003 16:43:36.154112 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb8803f7-c116-4175-b653-4b451e1da26d","Type":"ContainerStarted","Data":"3c3fbe34545f10996e3b7700b35eb0c4743eb42323a910487b6f2b9afae26aaf"} Oct 03 16:43:36 crc kubenswrapper[4744]: I1003 16:43:36.180971 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.180934604 podStartE2EDuration="6.180934604s" podCreationTimestamp="2025-10-03 16:43:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:36.16620679 +0000 UTC m=+1142.446082686" watchObservedRunningTime="2025-10-03 16:43:36.180934604 +0000 UTC m=+1142.460810500" Oct 03 16:43:37 crc kubenswrapper[4744]: I1003 16:43:37.166591 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb8803f7-c116-4175-b653-4b451e1da26d","Type":"ContainerStarted","Data":"44bc9bc5472efb50fcdd2da3dcd56942d91f2c82a55c8013f083377f384f0b06"} Oct 03 16:43:37 crc kubenswrapper[4744]: I1003 16:43:37.191432 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.191406729 podStartE2EDuration="6.191406729s" podCreationTimestamp="2025-10-03 16:43:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:43:37.187356128 +0000 UTC m=+1143.467232024" watchObservedRunningTime="2025-10-03 16:43:37.191406729 +0000 UTC m=+1143.471282625" Oct 03 16:43:39 crc kubenswrapper[4744]: I1003 16:43:39.206430 4744 generic.go:334] "Generic (PLEG): container finished" podID="245dc938-9219-43e5-8e69-10e033862451" containerID="0802af45ac185ed4e7848dce58259daaefb0b25cf43bdc9fae87eebd75be34ed" exitCode=0 Oct 03 16:43:39 crc kubenswrapper[4744]: I1003 16:43:39.206528 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9ftfh" event={"ID":"245dc938-9219-43e5-8e69-10e033862451","Type":"ContainerDied","Data":"0802af45ac185ed4e7848dce58259daaefb0b25cf43bdc9fae87eebd75be34ed"} Oct 03 16:43:39 crc kubenswrapper[4744]: I1003 16:43:39.377781 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:43:41 crc kubenswrapper[4744]: I1003 16:43:41.232771 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 16:43:41 crc kubenswrapper[4744]: I1003 16:43:41.234108 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 16:43:41 crc kubenswrapper[4744]: I1003 16:43:41.299241 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 16:43:41 crc kubenswrapper[4744]: I1003 16:43:41.373272 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 16:43:42 crc kubenswrapper[4744]: I1003 16:43:42.159149 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:42 crc kubenswrapper[4744]: I1003 16:43:42.159414 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:42 crc kubenswrapper[4744]: I1003 16:43:42.209925 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:42 crc kubenswrapper[4744]: I1003 16:43:42.214710 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:42 crc kubenswrapper[4744]: I1003 16:43:42.254453 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:42 crc kubenswrapper[4744]: I1003 16:43:42.254484 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 16:43:42 crc kubenswrapper[4744]: I1003 16:43:42.254495 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:42 crc kubenswrapper[4744]: I1003 16:43:42.254517 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 16:43:44 crc kubenswrapper[4744]: I1003 16:43:44.276138 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 16:43:44 crc kubenswrapper[4744]: I1003 16:43:44.276163 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 16:43:44 crc kubenswrapper[4744]: I1003 16:43:44.276985 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 16:43:44 crc kubenswrapper[4744]: I1003 16:43:44.276996 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 16:43:44 crc kubenswrapper[4744]: I1003 16:43:44.917279 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 16:43:45 crc kubenswrapper[4744]: I1003 16:43:45.082543 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:45 crc kubenswrapper[4744]: I1003 16:43:45.084266 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 16:43:45 crc kubenswrapper[4744]: I1003 16:43:45.296504 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 16:43:45 crc kubenswrapper[4744]: I1003 16:43:45.338029 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.018257 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7978488848-g4fcf" podUID="874834f7-8c7c-422d-9ec1-822d591f6d75" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.108267 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-85d8c4f7b-9pn28" podUID="39c711a0-3cea-4091-94a0-ec9df6e6a2d8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.605354 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.779050 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-fernet-keys\") pod \"245dc938-9219-43e5-8e69-10e033862451\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.779117 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-credential-keys\") pod \"245dc938-9219-43e5-8e69-10e033862451\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.779190 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wqvw\" (UniqueName: \"kubernetes.io/projected/245dc938-9219-43e5-8e69-10e033862451-kube-api-access-2wqvw\") pod \"245dc938-9219-43e5-8e69-10e033862451\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.779211 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-scripts\") pod \"245dc938-9219-43e5-8e69-10e033862451\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.779228 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-config-data\") pod \"245dc938-9219-43e5-8e69-10e033862451\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.779248 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-combined-ca-bundle\") pod \"245dc938-9219-43e5-8e69-10e033862451\" (UID: \"245dc938-9219-43e5-8e69-10e033862451\") " Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.787896 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "245dc938-9219-43e5-8e69-10e033862451" (UID: "245dc938-9219-43e5-8e69-10e033862451"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.789703 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-scripts" (OuterVolumeSpecName: "scripts") pod "245dc938-9219-43e5-8e69-10e033862451" (UID: "245dc938-9219-43e5-8e69-10e033862451"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.797450 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/245dc938-9219-43e5-8e69-10e033862451-kube-api-access-2wqvw" (OuterVolumeSpecName: "kube-api-access-2wqvw") pod "245dc938-9219-43e5-8e69-10e033862451" (UID: "245dc938-9219-43e5-8e69-10e033862451"). InnerVolumeSpecName "kube-api-access-2wqvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.799403 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "245dc938-9219-43e5-8e69-10e033862451" (UID: "245dc938-9219-43e5-8e69-10e033862451"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.850664 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "245dc938-9219-43e5-8e69-10e033862451" (UID: "245dc938-9219-43e5-8e69-10e033862451"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.881714 4744 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.881742 4744 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.881756 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wqvw\" (UniqueName: \"kubernetes.io/projected/245dc938-9219-43e5-8e69-10e033862451-kube-api-access-2wqvw\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.881764 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.881773 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.900683 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-config-data" (OuterVolumeSpecName: "config-data") pod "245dc938-9219-43e5-8e69-10e033862451" (UID: "245dc938-9219-43e5-8e69-10e033862451"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:43:46 crc kubenswrapper[4744]: I1003 16:43:46.983379 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/245dc938-9219-43e5-8e69-10e033862451-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.323413 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9ftfh" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.323338 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9ftfh" event={"ID":"245dc938-9219-43e5-8e69-10e033862451","Type":"ContainerDied","Data":"dee8d427f80c0aa195cab8dc7f68e0984ec3b09e2ac8660402c5ed9b553cd581"} Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.323530 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dee8d427f80c0aa195cab8dc7f68e0984ec3b09e2ac8660402c5ed9b553cd581" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.707519 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6589b6d4db-rbbtp"] Oct 03 16:43:47 crc kubenswrapper[4744]: E1003 16:43:47.707905 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="245dc938-9219-43e5-8e69-10e033862451" containerName="keystone-bootstrap" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.707924 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="245dc938-9219-43e5-8e69-10e033862451" containerName="keystone-bootstrap" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.708182 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="245dc938-9219-43e5-8e69-10e033862451" containerName="keystone-bootstrap" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.715082 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.717016 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.717591 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.718566 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-2z2zg" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.718688 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.719866 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6589b6d4db-rbbtp"] Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.721436 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.721754 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.799840 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-public-tls-certs\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.799892 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-config-data\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.799922 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-combined-ca-bundle\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.799949 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-scripts\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.799979 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-fernet-keys\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.799998 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-credential-keys\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.800024 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-948mq\" (UniqueName: \"kubernetes.io/projected/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-kube-api-access-948mq\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.800040 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-internal-tls-certs\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.902103 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-combined-ca-bundle\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.902164 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-scripts\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.902219 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-fernet-keys\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.902247 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-credential-keys\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.902303 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-948mq\" (UniqueName: \"kubernetes.io/projected/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-kube-api-access-948mq\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.902323 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-internal-tls-certs\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.902434 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-public-tls-certs\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.902476 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-config-data\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.914041 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-credential-keys\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.916431 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-combined-ca-bundle\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.916455 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-internal-tls-certs\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.918814 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-public-tls-certs\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.919090 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-scripts\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.920114 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-948mq\" (UniqueName: \"kubernetes.io/projected/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-kube-api-access-948mq\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.924169 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-fernet-keys\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:47 crc kubenswrapper[4744]: I1003 16:43:47.927053 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e8f7c20-0249-41e8-9cc6-915cf0762b3a-config-data\") pod \"keystone-6589b6d4db-rbbtp\" (UID: \"6e8f7c20-0249-41e8-9cc6-915cf0762b3a\") " pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:48 crc kubenswrapper[4744]: I1003 16:43:48.033202 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:43:56 crc kubenswrapper[4744]: I1003 16:43:56.017676 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7978488848-g4fcf" podUID="874834f7-8c7c-422d-9ec1-822d591f6d75" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Oct 03 16:43:56 crc kubenswrapper[4744]: I1003 16:43:56.430238 4744 generic.go:334] "Generic (PLEG): container finished" podID="3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7" containerID="6286e9ec39d254af81ef9ab41125f9cdbd2ce065a205572c2bbe1eb4f3a4bd0f" exitCode=0 Oct 03 16:43:56 crc kubenswrapper[4744]: I1003 16:43:56.430282 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-b88bn" event={"ID":"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7","Type":"ContainerDied","Data":"6286e9ec39d254af81ef9ab41125f9cdbd2ce065a205572c2bbe1eb4f3a4bd0f"} Oct 03 16:43:58 crc kubenswrapper[4744]: I1003 16:43:58.052158 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:59 crc kubenswrapper[4744]: I1003 16:43:59.826444 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-85d8c4f7b-9pn28" Oct 03 16:43:59 crc kubenswrapper[4744]: I1003 16:43:59.901632 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7978488848-g4fcf"] Oct 03 16:43:59 crc kubenswrapper[4744]: I1003 16:43:59.902396 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7978488848-g4fcf" podUID="874834f7-8c7c-422d-9ec1-822d591f6d75" containerName="horizon" containerID="cri-o://3fd2733e66910ca5adf8f373718198684ccaa21a7e01b73dbd60efbc2cae7e6c" gracePeriod=30 Oct 03 16:43:59 crc kubenswrapper[4744]: I1003 16:43:59.902300 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7978488848-g4fcf" podUID="874834f7-8c7c-422d-9ec1-822d591f6d75" containerName="horizon-log" containerID="cri-o://6ea220f0d199dc62a2062ec5f75b92116abceb965c5ab92e91a2b31c6d396809" gracePeriod=30 Oct 03 16:44:00 crc kubenswrapper[4744]: I1003 16:44:00.478433 4744 generic.go:334] "Generic (PLEG): container finished" podID="874834f7-8c7c-422d-9ec1-822d591f6d75" containerID="3fd2733e66910ca5adf8f373718198684ccaa21a7e01b73dbd60efbc2cae7e6c" exitCode=0 Oct 03 16:44:00 crc kubenswrapper[4744]: I1003 16:44:00.478523 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7978488848-g4fcf" event={"ID":"874834f7-8c7c-422d-9ec1-822d591f6d75","Type":"ContainerDied","Data":"3fd2733e66910ca5adf8f373718198684ccaa21a7e01b73dbd60efbc2cae7e6c"} Oct 03 16:44:00 crc kubenswrapper[4744]: E1003 16:44:00.495963 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 03 16:44:00 crc kubenswrapper[4744]: E1003 16:44:00.496127 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wqlt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-6k6b8_openstack(dc8def8d-6149-4797-b453-103ec15579f6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 16:44:00 crc kubenswrapper[4744]: E1003 16:44:00.497362 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-6k6b8" podUID="dc8def8d-6149-4797-b453-103ec15579f6" Oct 03 16:44:01 crc kubenswrapper[4744]: E1003 16:44:01.078818 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-manila-api:current-podified" Oct 03 16:44:01 crc kubenswrapper[4744]: E1003 16:44:01.078997 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manila-db-sync,Image:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,Command:[/bin/bash],Args:[-c sleep 0 && /usr/bin/manila-manage --config-dir /etc/manila/manila.conf.d db sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:job-config-data,ReadOnly:true,MountPath:/etc/manila/manila.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-58hb9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42429,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42429,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-db-sync-8bj8c_openstack(df7dff0d-99e0-4e98-b752-9096d24e314b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 16:44:01 crc kubenswrapper[4744]: E1003 16:44:01.081313 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manila-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/manila-db-sync-8bj8c" podUID="df7dff0d-99e0-4e98-b752-9096d24e314b" Oct 03 16:44:01 crc kubenswrapper[4744]: I1003 16:44:01.189761 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-b88bn" Oct 03 16:44:01 crc kubenswrapper[4744]: I1003 16:44:01.350314 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-config\") pod \"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7\" (UID: \"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7\") " Oct 03 16:44:01 crc kubenswrapper[4744]: I1003 16:44:01.350539 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lplm\" (UniqueName: \"kubernetes.io/projected/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-kube-api-access-7lplm\") pod \"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7\" (UID: \"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7\") " Oct 03 16:44:01 crc kubenswrapper[4744]: I1003 16:44:01.350585 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-combined-ca-bundle\") pod \"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7\" (UID: \"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7\") " Oct 03 16:44:01 crc kubenswrapper[4744]: I1003 16:44:01.355159 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-kube-api-access-7lplm" (OuterVolumeSpecName: "kube-api-access-7lplm") pod "3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7" (UID: "3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7"). InnerVolumeSpecName "kube-api-access-7lplm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:01 crc kubenswrapper[4744]: I1003 16:44:01.376665 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-config" (OuterVolumeSpecName: "config") pod "3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7" (UID: "3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:01 crc kubenswrapper[4744]: I1003 16:44:01.377058 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7" (UID: "3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:01 crc kubenswrapper[4744]: I1003 16:44:01.453098 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lplm\" (UniqueName: \"kubernetes.io/projected/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-kube-api-access-7lplm\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:01 crc kubenswrapper[4744]: I1003 16:44:01.454151 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:01 crc kubenswrapper[4744]: I1003 16:44:01.454164 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:01 crc kubenswrapper[4744]: I1003 16:44:01.488359 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-b88bn" event={"ID":"3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7","Type":"ContainerDied","Data":"c7ab859003cda6ba15c877d5259b1371534c24a3c3911044303c6a061b1ec6c4"} Oct 03 16:44:01 crc kubenswrapper[4744]: I1003 16:44:01.488410 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7ab859003cda6ba15c877d5259b1371534c24a3c3911044303c6a061b1ec6c4" Oct 03 16:44:01 crc kubenswrapper[4744]: I1003 16:44:01.488427 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-b88bn" Oct 03 16:44:01 crc kubenswrapper[4744]: E1003 16:44:01.490367 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manila-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-manila-api:current-podified\\\"\"" pod="openstack/manila-db-sync-8bj8c" podUID="df7dff0d-99e0-4e98-b752-9096d24e314b" Oct 03 16:44:01 crc kubenswrapper[4744]: E1003 16:44:01.493002 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-6k6b8" podUID="dc8def8d-6149-4797-b453-103ec15579f6" Oct 03 16:44:02 crc kubenswrapper[4744]: E1003 16:44:02.269389 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 03 16:44:02 crc kubenswrapper[4744]: E1003 16:44:02.269902 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-99bm5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-krczc_openstack(a8054599-30d8-42e0-bf45-2a8758b10486): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 16:44:02 crc kubenswrapper[4744]: E1003 16:44:02.271269 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-krczc" podUID="a8054599-30d8-42e0-bf45-2a8758b10486" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.354316 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-6bfh4"] Oct 03 16:44:02 crc kubenswrapper[4744]: E1003 16:44:02.355019 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7" containerName="neutron-db-sync" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.355032 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7" containerName="neutron-db-sync" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.355199 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7" containerName="neutron-db-sync" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.367644 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-6bfh4"] Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.367762 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.478737 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6fd7fcfb7b-kbrm2"] Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.480725 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.482271 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-config\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.482328 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.482377 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.482411 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcjtd\" (UniqueName: \"kubernetes.io/projected/4482beb6-50d9-4c37-8b51-927dbacbc455-kube-api-access-rcjtd\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.482451 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.482532 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.491999 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-gql6j" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.492180 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.492682 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.492796 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.510771 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6fd7fcfb7b-kbrm2"] Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.543536 4744 generic.go:334] "Generic (PLEG): container finished" podID="ecb56805-c2cc-4736-b5bc-84440acb7fa6" containerID="d9e099f2d91d174ed4080ff290be3a0e907e1d6ce44212fba5bdea7f9faf6350" exitCode=137 Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.543571 4744 generic.go:334] "Generic (PLEG): container finished" podID="ecb56805-c2cc-4736-b5bc-84440acb7fa6" containerID="dc4393a029969f5d1dded83e80cefa58f4efd64de2309c8dab88639cb4b8bce6" exitCode=137 Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.543617 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79c99f65c-jjfnc" event={"ID":"ecb56805-c2cc-4736-b5bc-84440acb7fa6","Type":"ContainerDied","Data":"d9e099f2d91d174ed4080ff290be3a0e907e1d6ce44212fba5bdea7f9faf6350"} Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.543640 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79c99f65c-jjfnc" event={"ID":"ecb56805-c2cc-4736-b5bc-84440acb7fa6","Type":"ContainerDied","Data":"dc4393a029969f5d1dded83e80cefa58f4efd64de2309c8dab88639cb4b8bce6"} Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.563126 4744 generic.go:334] "Generic (PLEG): container finished" podID="bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" containerID="25891acef34e96920974d576db823a20cd67c230f770dbc5f5e95ef8ee95fc40" exitCode=137 Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.563175 4744 generic.go:334] "Generic (PLEG): container finished" podID="bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" containerID="49414963259aa5d89aa752ae2d0e34de99d22af2fd18f7e139dd5e7d6c1886af" exitCode=137 Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.564240 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dc6d57fbc-4s82v" event={"ID":"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc","Type":"ContainerDied","Data":"25891acef34e96920974d576db823a20cd67c230f770dbc5f5e95ef8ee95fc40"} Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.564293 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dc6d57fbc-4s82v" event={"ID":"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc","Type":"ContainerDied","Data":"49414963259aa5d89aa752ae2d0e34de99d22af2fd18f7e139dd5e7d6c1886af"} Oct 03 16:44:02 crc kubenswrapper[4744]: E1003 16:44:02.575800 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-krczc" podUID="a8054599-30d8-42e0-bf45-2a8758b10486" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.587461 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-config\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.592891 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.593014 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-httpd-config\") pod \"neutron-6fd7fcfb7b-kbrm2\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.593072 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.593110 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcjtd\" (UniqueName: \"kubernetes.io/projected/4482beb6-50d9-4c37-8b51-927dbacbc455-kube-api-access-rcjtd\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.593158 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-combined-ca-bundle\") pod \"neutron-6fd7fcfb7b-kbrm2\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.595725 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-ovndb-tls-certs\") pod \"neutron-6fd7fcfb7b-kbrm2\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.595796 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.595866 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-config\") pod \"neutron-6fd7fcfb7b-kbrm2\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.595969 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.596037 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqnrn\" (UniqueName: \"kubernetes.io/projected/21f7869b-e2f2-4e15-861c-9c3df31aa03e-kube-api-access-xqnrn\") pod \"neutron-6fd7fcfb7b-kbrm2\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.592778 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-config\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.594632 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.595289 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.605936 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.606918 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.618538 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcjtd\" (UniqueName: \"kubernetes.io/projected/4482beb6-50d9-4c37-8b51-927dbacbc455-kube-api-access-rcjtd\") pod \"dnsmasq-dns-84b966f6c9-6bfh4\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.713166 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-config\") pod \"neutron-6fd7fcfb7b-kbrm2\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.714450 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqnrn\" (UniqueName: \"kubernetes.io/projected/21f7869b-e2f2-4e15-861c-9c3df31aa03e-kube-api-access-xqnrn\") pod \"neutron-6fd7fcfb7b-kbrm2\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.715486 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-httpd-config\") pod \"neutron-6fd7fcfb7b-kbrm2\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.716141 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-combined-ca-bundle\") pod \"neutron-6fd7fcfb7b-kbrm2\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.716220 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-ovndb-tls-certs\") pod \"neutron-6fd7fcfb7b-kbrm2\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.727189 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-config\") pod \"neutron-6fd7fcfb7b-kbrm2\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.727764 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-ovndb-tls-certs\") pod \"neutron-6fd7fcfb7b-kbrm2\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.733508 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-combined-ca-bundle\") pod \"neutron-6fd7fcfb7b-kbrm2\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.736074 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-httpd-config\") pod \"neutron-6fd7fcfb7b-kbrm2\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.737074 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqnrn\" (UniqueName: \"kubernetes.io/projected/21f7869b-e2f2-4e15-861c-9c3df31aa03e-kube-api-access-xqnrn\") pod \"neutron-6fd7fcfb7b-kbrm2\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.794717 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.828662 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.847846 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:44:02 crc kubenswrapper[4744]: I1003 16:44:02.912840 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.028069 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-logs\") pod \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.028484 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcpzb\" (UniqueName: \"kubernetes.io/projected/ecb56805-c2cc-4736-b5bc-84440acb7fa6-kube-api-access-mcpzb\") pod \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.028552 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-config-data\") pod \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.028581 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9b9xr\" (UniqueName: \"kubernetes.io/projected/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-kube-api-access-9b9xr\") pod \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.028655 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecb56805-c2cc-4736-b5bc-84440acb7fa6-logs\") pod \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.028680 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-horizon-secret-key\") pod \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.028744 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ecb56805-c2cc-4736-b5bc-84440acb7fa6-config-data\") pod \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.028772 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecb56805-c2cc-4736-b5bc-84440acb7fa6-scripts\") pod \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.028828 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ecb56805-c2cc-4736-b5bc-84440acb7fa6-horizon-secret-key\") pod \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\" (UID: \"ecb56805-c2cc-4736-b5bc-84440acb7fa6\") " Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.028853 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-scripts\") pod \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\" (UID: \"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc\") " Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.029716 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-logs" (OuterVolumeSpecName: "logs") pod "bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" (UID: "bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.033140 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecb56805-c2cc-4736-b5bc-84440acb7fa6-logs" (OuterVolumeSpecName: "logs") pod "ecb56805-c2cc-4736-b5bc-84440acb7fa6" (UID: "ecb56805-c2cc-4736-b5bc-84440acb7fa6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.037159 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" (UID: "bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.044135 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-kube-api-access-9b9xr" (OuterVolumeSpecName: "kube-api-access-9b9xr") pod "bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" (UID: "bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc"). InnerVolumeSpecName "kube-api-access-9b9xr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.059021 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecb56805-c2cc-4736-b5bc-84440acb7fa6-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ecb56805-c2cc-4736-b5bc-84440acb7fa6" (UID: "ecb56805-c2cc-4736-b5bc-84440acb7fa6"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.063539 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-config-data" (OuterVolumeSpecName: "config-data") pod "bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" (UID: "bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.068945 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecb56805-c2cc-4736-b5bc-84440acb7fa6-kube-api-access-mcpzb" (OuterVolumeSpecName: "kube-api-access-mcpzb") pod "ecb56805-c2cc-4736-b5bc-84440acb7fa6" (UID: "ecb56805-c2cc-4736-b5bc-84440acb7fa6"). InnerVolumeSpecName "kube-api-access-mcpzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.074619 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6589b6d4db-rbbtp"] Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.097104 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecb56805-c2cc-4736-b5bc-84440acb7fa6-scripts" (OuterVolumeSpecName: "scripts") pod "ecb56805-c2cc-4736-b5bc-84440acb7fa6" (UID: "ecb56805-c2cc-4736-b5bc-84440acb7fa6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.106417 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-scripts" (OuterVolumeSpecName: "scripts") pod "bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" (UID: "bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.106630 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecb56805-c2cc-4736-b5bc-84440acb7fa6-config-data" (OuterVolumeSpecName: "config-data") pod "ecb56805-c2cc-4736-b5bc-84440acb7fa6" (UID: "ecb56805-c2cc-4736-b5bc-84440acb7fa6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:03 crc kubenswrapper[4744]: W1003 16:44:03.107765 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e8f7c20_0249_41e8_9cc6_915cf0762b3a.slice/crio-d76274ef3537185b45d516740bba987c7fe411b908285cc5f1fa9e4ca36c910f WatchSource:0}: Error finding container d76274ef3537185b45d516740bba987c7fe411b908285cc5f1fa9e4ca36c910f: Status 404 returned error can't find the container with id d76274ef3537185b45d516740bba987c7fe411b908285cc5f1fa9e4ca36c910f Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.130443 4744 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ecb56805-c2cc-4736-b5bc-84440acb7fa6-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.130469 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.130480 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.130490 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcpzb\" (UniqueName: \"kubernetes.io/projected/ecb56805-c2cc-4736-b5bc-84440acb7fa6-kube-api-access-mcpzb\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.130512 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.130521 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9b9xr\" (UniqueName: \"kubernetes.io/projected/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-kube-api-access-9b9xr\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.130530 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecb56805-c2cc-4736-b5bc-84440acb7fa6-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.130538 4744 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.130546 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ecb56805-c2cc-4736-b5bc-84440acb7fa6-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.130553 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecb56805-c2cc-4736-b5bc-84440acb7fa6-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:03 crc kubenswrapper[4744]: W1003 16:44:03.388335 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4482beb6_50d9_4c37_8b51_927dbacbc455.slice/crio-074afbc4c4fff049b90bb08afdf33f4a9d95b28878621ecb42443352e4586732 WatchSource:0}: Error finding container 074afbc4c4fff049b90bb08afdf33f4a9d95b28878621ecb42443352e4586732: Status 404 returned error can't find the container with id 074afbc4c4fff049b90bb08afdf33f4a9d95b28878621ecb42443352e4586732 Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.389213 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-6bfh4"] Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.544926 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.645676 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d991f026-62e5-4515-924d-ae7267cab11f-horizon-secret-key\") pod \"d991f026-62e5-4515-924d-ae7267cab11f\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.646063 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmxsd\" (UniqueName: \"kubernetes.io/projected/d991f026-62e5-4515-924d-ae7267cab11f-kube-api-access-zmxsd\") pod \"d991f026-62e5-4515-924d-ae7267cab11f\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.646201 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d991f026-62e5-4515-924d-ae7267cab11f-logs\") pod \"d991f026-62e5-4515-924d-ae7267cab11f\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.646364 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d991f026-62e5-4515-924d-ae7267cab11f-config-data\") pod \"d991f026-62e5-4515-924d-ae7267cab11f\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.646481 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d991f026-62e5-4515-924d-ae7267cab11f-scripts\") pod \"d991f026-62e5-4515-924d-ae7267cab11f\" (UID: \"d991f026-62e5-4515-924d-ae7267cab11f\") " Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.652176 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d991f026-62e5-4515-924d-ae7267cab11f-logs" (OuterVolumeSpecName: "logs") pod "d991f026-62e5-4515-924d-ae7267cab11f" (UID: "d991f026-62e5-4515-924d-ae7267cab11f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.662738 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d991f026-62e5-4515-924d-ae7267cab11f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d991f026-62e5-4515-924d-ae7267cab11f" (UID: "d991f026-62e5-4515-924d-ae7267cab11f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.671011 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d991f026-62e5-4515-924d-ae7267cab11f-kube-api-access-zmxsd" (OuterVolumeSpecName: "kube-api-access-zmxsd") pod "d991f026-62e5-4515-924d-ae7267cab11f" (UID: "d991f026-62e5-4515-924d-ae7267cab11f"). InnerVolumeSpecName "kube-api-access-zmxsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.697330 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-79c99f65c-jjfnc" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.699246 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-79c99f65c-jjfnc" event={"ID":"ecb56805-c2cc-4736-b5bc-84440acb7fa6","Type":"ContainerDied","Data":"64b7897d9087f1f10340bf44291a6b2e447d33e43b3eaccc10bd6ae15875b449"} Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.699313 4744 scope.go:117] "RemoveContainer" containerID="d9e099f2d91d174ed4080ff290be3a0e907e1d6ce44212fba5bdea7f9faf6350" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.705722 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6589b6d4db-rbbtp" event={"ID":"6e8f7c20-0249-41e8-9cc6-915cf0762b3a","Type":"ContainerStarted","Data":"e92e0d52bfb3bf0e63e7f39768e840996592eddbfe30dff8df7ba6eb79c74a38"} Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.705764 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6589b6d4db-rbbtp" event={"ID":"6e8f7c20-0249-41e8-9cc6-915cf0762b3a","Type":"ContainerStarted","Data":"d76274ef3537185b45d516740bba987c7fe411b908285cc5f1fa9e4ca36c910f"} Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.706455 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.733802 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dc6d57fbc-4s82v" event={"ID":"bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc","Type":"ContainerDied","Data":"e6299623c2bfc959a73601967bea901378c7bc147dc129ded91ff9aae570dae1"} Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.733940 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-dc6d57fbc-4s82v" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.734298 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6fd7fcfb7b-kbrm2"] Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.739182 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6589b6d4db-rbbtp" podStartSLOduration=16.739158617 podStartE2EDuration="16.739158617s" podCreationTimestamp="2025-10-03 16:43:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:03.732790822 +0000 UTC m=+1170.012666718" watchObservedRunningTime="2025-10-03 16:44:03.739158617 +0000 UTC m=+1170.019034503" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.752015 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d991f026-62e5-4515-924d-ae7267cab11f-scripts" (OuterVolumeSpecName: "scripts") pod "d991f026-62e5-4515-924d-ae7267cab11f" (UID: "d991f026-62e5-4515-924d-ae7267cab11f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.752745 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmxsd\" (UniqueName: \"kubernetes.io/projected/d991f026-62e5-4515-924d-ae7267cab11f-kube-api-access-zmxsd\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.752768 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d991f026-62e5-4515-924d-ae7267cab11f-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.752778 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d991f026-62e5-4515-924d-ae7267cab11f-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.752804 4744 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d991f026-62e5-4515-924d-ae7267cab11f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.761748 4744 generic.go:334] "Generic (PLEG): container finished" podID="d991f026-62e5-4515-924d-ae7267cab11f" containerID="1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3" exitCode=137 Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.761788 4744 generic.go:334] "Generic (PLEG): container finished" podID="d991f026-62e5-4515-924d-ae7267cab11f" containerID="e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7" exitCode=137 Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.761836 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75bcc4d5c-s79cg" event={"ID":"d991f026-62e5-4515-924d-ae7267cab11f","Type":"ContainerDied","Data":"1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3"} Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.761867 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75bcc4d5c-s79cg" event={"ID":"d991f026-62e5-4515-924d-ae7267cab11f","Type":"ContainerDied","Data":"e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7"} Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.761880 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75bcc4d5c-s79cg" event={"ID":"d991f026-62e5-4515-924d-ae7267cab11f","Type":"ContainerDied","Data":"61f5c36bba91ff88181f6563fedd5e86a33d4a733ff0dd6a9b10ebfd4202c48c"} Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.761949 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75bcc4d5c-s79cg" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.763541 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d991f026-62e5-4515-924d-ae7267cab11f-config-data" (OuterVolumeSpecName: "config-data") pod "d991f026-62e5-4515-924d-ae7267cab11f" (UID: "d991f026-62e5-4515-924d-ae7267cab11f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.769562 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-79c99f65c-jjfnc"] Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.773789 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qbz46" event={"ID":"7acfc27b-28b7-422c-a933-8dd2a4a99c68","Type":"ContainerStarted","Data":"0fe5a6ae0fcdb499668ae953e15d63f01addf862e36032e8e66e70b365a4ba0c"} Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.781637 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-79c99f65c-jjfnc"] Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.788994 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-dc6d57fbc-4s82v"] Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.799902 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4","Type":"ContainerStarted","Data":"e56c8cf23f7e569c360c631b099ce8e5cd11460568423c06b51f9baaa1b6bac5"} Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.800300 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-dc6d57fbc-4s82v"] Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.802038 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" event={"ID":"4482beb6-50d9-4c37-8b51-927dbacbc455","Type":"ContainerStarted","Data":"074afbc4c4fff049b90bb08afdf33f4a9d95b28878621ecb42443352e4586732"} Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.810982 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-qbz46" podStartSLOduration=4.145949009 podStartE2EDuration="49.810966494s" podCreationTimestamp="2025-10-03 16:43:14 +0000 UTC" firstStartedPulling="2025-10-03 16:43:15.397786822 +0000 UTC m=+1121.677662718" lastFinishedPulling="2025-10-03 16:44:01.062804307 +0000 UTC m=+1167.342680203" observedRunningTime="2025-10-03 16:44:03.804751834 +0000 UTC m=+1170.084627730" watchObservedRunningTime="2025-10-03 16:44:03.810966494 +0000 UTC m=+1170.090842380" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.854617 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d991f026-62e5-4515-924d-ae7267cab11f-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.927830 4744 scope.go:117] "RemoveContainer" containerID="dc4393a029969f5d1dded83e80cefa58f4efd64de2309c8dab88639cb4b8bce6" Oct 03 16:44:03 crc kubenswrapper[4744]: I1003 16:44:03.952675 4744 scope.go:117] "RemoveContainer" containerID="25891acef34e96920974d576db823a20cd67c230f770dbc5f5e95ef8ee95fc40" Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.110353 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-75bcc4d5c-s79cg"] Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.132603 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-75bcc4d5c-s79cg"] Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.182647 4744 scope.go:117] "RemoveContainer" containerID="49414963259aa5d89aa752ae2d0e34de99d22af2fd18f7e139dd5e7d6c1886af" Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.264651 4744 scope.go:117] "RemoveContainer" containerID="1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3" Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.505993 4744 scope.go:117] "RemoveContainer" containerID="e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7" Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.529299 4744 scope.go:117] "RemoveContainer" containerID="1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3" Oct 03 16:44:04 crc kubenswrapper[4744]: E1003 16:44:04.529798 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3\": container with ID starting with 1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3 not found: ID does not exist" containerID="1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3" Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.529843 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3"} err="failed to get container status \"1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3\": rpc error: code = NotFound desc = could not find container \"1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3\": container with ID starting with 1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3 not found: ID does not exist" Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.529869 4744 scope.go:117] "RemoveContainer" containerID="e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7" Oct 03 16:44:04 crc kubenswrapper[4744]: E1003 16:44:04.530233 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7\": container with ID starting with e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7 not found: ID does not exist" containerID="e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7" Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.530309 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7"} err="failed to get container status \"e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7\": rpc error: code = NotFound desc = could not find container \"e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7\": container with ID starting with e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7 not found: ID does not exist" Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.530362 4744 scope.go:117] "RemoveContainer" containerID="1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3" Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.530720 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3"} err="failed to get container status \"1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3\": rpc error: code = NotFound desc = could not find container \"1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3\": container with ID starting with 1e5ec3576cdf09f65a782abf06be0f372cfb05cf47585bda599368fbdb2b88e3 not found: ID does not exist" Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.530749 4744 scope.go:117] "RemoveContainer" containerID="e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7" Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.531123 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7"} err="failed to get container status \"e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7\": rpc error: code = NotFound desc = could not find container \"e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7\": container with ID starting with e9337f789a8786786fca66b34e155a4bb96919247359ea9b192591a287f2f2e7 not found: ID does not exist" Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.668069 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.668129 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.816378 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fd7fcfb7b-kbrm2" event={"ID":"21f7869b-e2f2-4e15-861c-9c3df31aa03e","Type":"ContainerStarted","Data":"d2cde1ba2ae3a2a6c2ad6e9d4f4fdef85ef6c41bd2856cd5e4623d1dff03487a"} Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.816683 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fd7fcfb7b-kbrm2" event={"ID":"21f7869b-e2f2-4e15-861c-9c3df31aa03e","Type":"ContainerStarted","Data":"bbc5a489ad2c5f3da6cc3beb9575ad8ad8298e98dda6a843c47746732300e126"} Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.818175 4744 generic.go:334] "Generic (PLEG): container finished" podID="4482beb6-50d9-4c37-8b51-927dbacbc455" containerID="3b70138207949fd8d39b629b71a2550352f9a5faa2d9ce32c6a3c59417f6993b" exitCode=0 Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.818283 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" event={"ID":"4482beb6-50d9-4c37-8b51-927dbacbc455","Type":"ContainerDied","Data":"3b70138207949fd8d39b629b71a2550352f9a5faa2d9ce32c6a3c59417f6993b"} Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.906047 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" path="/var/lib/kubelet/pods/bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc/volumes" Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.906771 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d991f026-62e5-4515-924d-ae7267cab11f" path="/var/lib/kubelet/pods/d991f026-62e5-4515-924d-ae7267cab11f/volumes" Oct 03 16:44:04 crc kubenswrapper[4744]: I1003 16:44:04.910169 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecb56805-c2cc-4736-b5bc-84440acb7fa6" path="/var/lib/kubelet/pods/ecb56805-c2cc-4736-b5bc-84440acb7fa6/volumes" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.467647 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6bc59588f-s74gr"] Oct 03 16:44:05 crc kubenswrapper[4744]: E1003 16:44:05.468023 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecb56805-c2cc-4736-b5bc-84440acb7fa6" containerName="horizon" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.468040 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecb56805-c2cc-4736-b5bc-84440acb7fa6" containerName="horizon" Oct 03 16:44:05 crc kubenswrapper[4744]: E1003 16:44:05.468059 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d991f026-62e5-4515-924d-ae7267cab11f" containerName="horizon-log" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.468066 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d991f026-62e5-4515-924d-ae7267cab11f" containerName="horizon-log" Oct 03 16:44:05 crc kubenswrapper[4744]: E1003 16:44:05.468075 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" containerName="horizon" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.468081 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" containerName="horizon" Oct 03 16:44:05 crc kubenswrapper[4744]: E1003 16:44:05.468087 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" containerName="horizon-log" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.468093 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" containerName="horizon-log" Oct 03 16:44:05 crc kubenswrapper[4744]: E1003 16:44:05.468105 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecb56805-c2cc-4736-b5bc-84440acb7fa6" containerName="horizon-log" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.468110 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecb56805-c2cc-4736-b5bc-84440acb7fa6" containerName="horizon-log" Oct 03 16:44:05 crc kubenswrapper[4744]: E1003 16:44:05.468126 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d991f026-62e5-4515-924d-ae7267cab11f" containerName="horizon" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.468132 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d991f026-62e5-4515-924d-ae7267cab11f" containerName="horizon" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.468286 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" containerName="horizon-log" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.468297 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf9b9eea-c5d4-4009-8f8d-c0a73648a8bc" containerName="horizon" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.468313 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d991f026-62e5-4515-924d-ae7267cab11f" containerName="horizon" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.468324 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecb56805-c2cc-4736-b5bc-84440acb7fa6" containerName="horizon" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.468336 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecb56805-c2cc-4736-b5bc-84440acb7fa6" containerName="horizon-log" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.468345 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d991f026-62e5-4515-924d-ae7267cab11f" containerName="horizon-log" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.469343 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.471695 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.473308 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.498598 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6bc59588f-s74gr"] Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.602053 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-public-tls-certs\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.604558 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj6wp\" (UniqueName: \"kubernetes.io/projected/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-kube-api-access-fj6wp\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.604742 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-ovndb-tls-certs\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.604873 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-combined-ca-bundle\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.604952 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-config\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.605039 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-internal-tls-certs\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.605261 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-httpd-config\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.707204 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-public-tls-certs\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.707265 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj6wp\" (UniqueName: \"kubernetes.io/projected/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-kube-api-access-fj6wp\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.707311 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-ovndb-tls-certs\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.707349 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-combined-ca-bundle\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.707369 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-config\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.707394 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-internal-tls-certs\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.707457 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-httpd-config\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.712401 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-httpd-config\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.712942 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-config\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.713802 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-internal-tls-certs\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.713802 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-public-tls-certs\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.714222 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-ovndb-tls-certs\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.715432 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-combined-ca-bundle\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.731625 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj6wp\" (UniqueName: \"kubernetes.io/projected/ec5a5a4c-4896-4629-be5d-2b96f2729eaf-kube-api-access-fj6wp\") pod \"neutron-6bc59588f-s74gr\" (UID: \"ec5a5a4c-4896-4629-be5d-2b96f2729eaf\") " pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:05 crc kubenswrapper[4744]: I1003 16:44:05.788046 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:06 crc kubenswrapper[4744]: I1003 16:44:06.387863 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6bc59588f-s74gr"] Oct 03 16:44:06 crc kubenswrapper[4744]: W1003 16:44:06.387912 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec5a5a4c_4896_4629_be5d_2b96f2729eaf.slice/crio-650eda69c186ba40b38036521bd9e2710aea996f6d14ede3c389635553ca2785 WatchSource:0}: Error finding container 650eda69c186ba40b38036521bd9e2710aea996f6d14ede3c389635553ca2785: Status 404 returned error can't find the container with id 650eda69c186ba40b38036521bd9e2710aea996f6d14ede3c389635553ca2785 Oct 03 16:44:06 crc kubenswrapper[4744]: I1003 16:44:06.853486 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bc59588f-s74gr" event={"ID":"ec5a5a4c-4896-4629-be5d-2b96f2729eaf","Type":"ContainerStarted","Data":"650eda69c186ba40b38036521bd9e2710aea996f6d14ede3c389635553ca2785"} Oct 03 16:44:08 crc kubenswrapper[4744]: I1003 16:44:08.875990 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" event={"ID":"4482beb6-50d9-4c37-8b51-927dbacbc455","Type":"ContainerStarted","Data":"ce53319130146b9d4474d482dcd8399ab7aac5d448defefe0fe4d9f643c1484d"} Oct 03 16:44:08 crc kubenswrapper[4744]: I1003 16:44:08.876554 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:08 crc kubenswrapper[4744]: I1003 16:44:08.879060 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bc59588f-s74gr" event={"ID":"ec5a5a4c-4896-4629-be5d-2b96f2729eaf","Type":"ContainerStarted","Data":"513eda01f462270a7a668d550b7054fce9ec2e68634fa3c3ae1be1d3bbf83337"} Oct 03 16:44:08 crc kubenswrapper[4744]: I1003 16:44:08.903390 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" podStartSLOduration=6.903375744 podStartE2EDuration="6.903375744s" podCreationTimestamp="2025-10-03 16:44:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:08.895983086 +0000 UTC m=+1175.175858982" watchObservedRunningTime="2025-10-03 16:44:08.903375744 +0000 UTC m=+1175.183251640" Oct 03 16:44:09 crc kubenswrapper[4744]: I1003 16:44:09.896275 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fd7fcfb7b-kbrm2" event={"ID":"21f7869b-e2f2-4e15-861c-9c3df31aa03e","Type":"ContainerStarted","Data":"7b0dcaeee0ce21d78370b7e5c3ab8b3d6209a881fa53dbe65f50bcc2438ff8c3"} Oct 03 16:44:09 crc kubenswrapper[4744]: I1003 16:44:09.896634 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:09 crc kubenswrapper[4744]: I1003 16:44:09.901040 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bc59588f-s74gr" event={"ID":"ec5a5a4c-4896-4629-be5d-2b96f2729eaf","Type":"ContainerStarted","Data":"594eaa3cc985348be87a55d991dce86a997e98cca201077d8610e291bfe6eeef"} Oct 03 16:44:09 crc kubenswrapper[4744]: I1003 16:44:09.924648 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6fd7fcfb7b-kbrm2" podStartSLOduration=7.9246251149999996 podStartE2EDuration="7.924625115s" podCreationTimestamp="2025-10-03 16:44:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:09.911704652 +0000 UTC m=+1176.191580548" watchObservedRunningTime="2025-10-03 16:44:09.924625115 +0000 UTC m=+1176.204501011" Oct 03 16:44:09 crc kubenswrapper[4744]: I1003 16:44:09.935967 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6bc59588f-s74gr" podStartSLOduration=4.935951832 podStartE2EDuration="4.935951832s" podCreationTimestamp="2025-10-03 16:44:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:09.929440524 +0000 UTC m=+1176.209316420" watchObservedRunningTime="2025-10-03 16:44:09.935951832 +0000 UTC m=+1176.215827728" Oct 03 16:44:10 crc kubenswrapper[4744]: I1003 16:44:10.920133 4744 generic.go:334] "Generic (PLEG): container finished" podID="7acfc27b-28b7-422c-a933-8dd2a4a99c68" containerID="0fe5a6ae0fcdb499668ae953e15d63f01addf862e36032e8e66e70b365a4ba0c" exitCode=0 Oct 03 16:44:10 crc kubenswrapper[4744]: I1003 16:44:10.920190 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qbz46" event={"ID":"7acfc27b-28b7-422c-a933-8dd2a4a99c68","Type":"ContainerDied","Data":"0fe5a6ae0fcdb499668ae953e15d63f01addf862e36032e8e66e70b365a4ba0c"} Oct 03 16:44:10 crc kubenswrapper[4744]: I1003 16:44:10.920380 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.076074 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qbz46" Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.229577 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7acfc27b-28b7-422c-a933-8dd2a4a99c68-logs\") pod \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.229670 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-combined-ca-bundle\") pod \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.229801 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5z6qc\" (UniqueName: \"kubernetes.io/projected/7acfc27b-28b7-422c-a933-8dd2a4a99c68-kube-api-access-5z6qc\") pod \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.229855 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-scripts\") pod \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.229877 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-config-data\") pod \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\" (UID: \"7acfc27b-28b7-422c-a933-8dd2a4a99c68\") " Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.230282 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7acfc27b-28b7-422c-a933-8dd2a4a99c68-logs" (OuterVolumeSpecName: "logs") pod "7acfc27b-28b7-422c-a933-8dd2a4a99c68" (UID: "7acfc27b-28b7-422c-a933-8dd2a4a99c68"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.236370 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-scripts" (OuterVolumeSpecName: "scripts") pod "7acfc27b-28b7-422c-a933-8dd2a4a99c68" (UID: "7acfc27b-28b7-422c-a933-8dd2a4a99c68"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.236476 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7acfc27b-28b7-422c-a933-8dd2a4a99c68-kube-api-access-5z6qc" (OuterVolumeSpecName: "kube-api-access-5z6qc") pod "7acfc27b-28b7-422c-a933-8dd2a4a99c68" (UID: "7acfc27b-28b7-422c-a933-8dd2a4a99c68"). InnerVolumeSpecName "kube-api-access-5z6qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.259456 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7acfc27b-28b7-422c-a933-8dd2a4a99c68" (UID: "7acfc27b-28b7-422c-a933-8dd2a4a99c68"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.261685 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-config-data" (OuterVolumeSpecName: "config-data") pod "7acfc27b-28b7-422c-a933-8dd2a4a99c68" (UID: "7acfc27b-28b7-422c-a933-8dd2a4a99c68"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.331450 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5z6qc\" (UniqueName: \"kubernetes.io/projected/7acfc27b-28b7-422c-a933-8dd2a4a99c68-kube-api-access-5z6qc\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.331500 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.331512 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.331522 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7acfc27b-28b7-422c-a933-8dd2a4a99c68-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.331532 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7acfc27b-28b7-422c-a933-8dd2a4a99c68-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.985692 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4","Type":"ContainerStarted","Data":"0b97444af620bd92a44ee61d76bf19333ee8ac3674e16796c07f625342ded3b3"} Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.986052 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.985772 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="ceilometer-central-agent" containerID="cri-o://c4f9d07a32023e6cae16d77eac1219e6d66fcd4e519e8880aed981dd3e712ebe" gracePeriod=30 Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.986135 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="proxy-httpd" containerID="cri-o://0b97444af620bd92a44ee61d76bf19333ee8ac3674e16796c07f625342ded3b3" gracePeriod=30 Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.986212 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="ceilometer-notification-agent" containerID="cri-o://7eb661e5353163eb93dc19558c766a2c4d32ad30ebec5298fa08a7861f1a24d1" gracePeriod=30 Oct 03 16:44:16 crc kubenswrapper[4744]: I1003 16:44:16.986249 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="sg-core" containerID="cri-o://e56c8cf23f7e569c360c631b099ce8e5cd11460568423c06b51f9baaa1b6bac5" gracePeriod=30 Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.002406 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6k6b8" event={"ID":"dc8def8d-6149-4797-b453-103ec15579f6","Type":"ContainerStarted","Data":"8c2b637e9fea89323ced9b6a7c16ff4a1499d858e2fb5fcb8ba4fb3b95f37182"} Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.005329 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qbz46" event={"ID":"7acfc27b-28b7-422c-a933-8dd2a4a99c68","Type":"ContainerDied","Data":"28e3444a85024e03ece9face876aab8a91e8b2967b4a712f64ce863e5052661b"} Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.005383 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28e3444a85024e03ece9face876aab8a91e8b2967b4a712f64ce863e5052661b" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.005421 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qbz46" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.016856 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.913555994 podStartE2EDuration="1m3.016839337s" podCreationTimestamp="2025-10-03 16:43:14 +0000 UTC" firstStartedPulling="2025-10-03 16:43:15.493733317 +0000 UTC m=+1121.773609213" lastFinishedPulling="2025-10-03 16:44:16.59701666 +0000 UTC m=+1182.876892556" observedRunningTime="2025-10-03 16:44:17.014993825 +0000 UTC m=+1183.294869721" watchObservedRunningTime="2025-10-03 16:44:17.016839337 +0000 UTC m=+1183.296715233" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.039868 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-6k6b8" podStartSLOduration=4.170080489 podStartE2EDuration="46.039844459s" podCreationTimestamp="2025-10-03 16:43:31 +0000 UTC" firstStartedPulling="2025-10-03 16:43:34.727860714 +0000 UTC m=+1141.007736610" lastFinishedPulling="2025-10-03 16:44:16.597624644 +0000 UTC m=+1182.877500580" observedRunningTime="2025-10-03 16:44:17.033535306 +0000 UTC m=+1183.313411202" watchObservedRunningTime="2025-10-03 16:44:17.039844459 +0000 UTC m=+1183.319720355" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.186305 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-774fb64f68-qgzkk"] Oct 03 16:44:17 crc kubenswrapper[4744]: E1003 16:44:17.186971 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7acfc27b-28b7-422c-a933-8dd2a4a99c68" containerName="placement-db-sync" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.186983 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7acfc27b-28b7-422c-a933-8dd2a4a99c68" containerName="placement-db-sync" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.187183 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7acfc27b-28b7-422c-a933-8dd2a4a99c68" containerName="placement-db-sync" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.188091 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.190157 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-749k2" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.190841 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.190848 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.191397 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.191611 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.203620 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-774fb64f68-qgzkk"] Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.354455 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-config-data\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.354718 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-combined-ca-bundle\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.354789 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-internal-tls-certs\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.354910 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-public-tls-certs\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.355115 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-logs\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.355164 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsb9g\" (UniqueName: \"kubernetes.io/projected/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-kube-api-access-xsb9g\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.355192 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-scripts\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.456944 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsb9g\" (UniqueName: \"kubernetes.io/projected/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-kube-api-access-xsb9g\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.457000 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-scripts\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.457048 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-config-data\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.457145 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-combined-ca-bundle\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.457162 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-internal-tls-certs\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.457194 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-public-tls-certs\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.457258 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-logs\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.457659 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-logs\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.463169 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-internal-tls-certs\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.463241 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-config-data\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.464271 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-public-tls-certs\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.466218 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-combined-ca-bundle\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.466788 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-scripts\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.498135 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsb9g\" (UniqueName: \"kubernetes.io/projected/0d2b9c37-6d79-420d-bc4e-3d5888b5d61e-kube-api-access-xsb9g\") pod \"placement-774fb64f68-qgzkk\" (UID: \"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e\") " pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.558160 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.801708 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.884128 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-ndwh8"] Oct 03 16:44:17 crc kubenswrapper[4744]: I1003 16:44:17.884346 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" podUID="7ca5eb4d-d2b1-4357-8b49-001763e93804" containerName="dnsmasq-dns" containerID="cri-o://42af22cf2a27fc349e7615656c18eb71e7b78bf252206da91e54c7bcca85a59a" gracePeriod=10 Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.030877 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-8bj8c" event={"ID":"df7dff0d-99e0-4e98-b752-9096d24e314b","Type":"ContainerStarted","Data":"e7656f6bbb2b5a8116c6daa66427fab681e8be77ee5904e4f595b06426e76e0e"} Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.040352 4744 generic.go:334] "Generic (PLEG): container finished" podID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerID="0b97444af620bd92a44ee61d76bf19333ee8ac3674e16796c07f625342ded3b3" exitCode=0 Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.040416 4744 generic.go:334] "Generic (PLEG): container finished" podID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerID="e56c8cf23f7e569c360c631b099ce8e5cd11460568423c06b51f9baaa1b6bac5" exitCode=2 Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.040427 4744 generic.go:334] "Generic (PLEG): container finished" podID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerID="c4f9d07a32023e6cae16d77eac1219e6d66fcd4e519e8880aed981dd3e712ebe" exitCode=0 Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.040476 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4","Type":"ContainerDied","Data":"0b97444af620bd92a44ee61d76bf19333ee8ac3674e16796c07f625342ded3b3"} Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.040536 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4","Type":"ContainerDied","Data":"e56c8cf23f7e569c360c631b099ce8e5cd11460568423c06b51f9baaa1b6bac5"} Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.040555 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4","Type":"ContainerDied","Data":"c4f9d07a32023e6cae16d77eac1219e6d66fcd4e519e8880aed981dd3e712ebe"} Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.073427 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-8bj8c" podStartSLOduration=5.193091429 podStartE2EDuration="47.073406869s" podCreationTimestamp="2025-10-03 16:43:31 +0000 UTC" firstStartedPulling="2025-10-03 16:43:34.71491123 +0000 UTC m=+1140.994787126" lastFinishedPulling="2025-10-03 16:44:16.59522667 +0000 UTC m=+1182.875102566" observedRunningTime="2025-10-03 16:44:18.055985754 +0000 UTC m=+1184.335861660" watchObservedRunningTime="2025-10-03 16:44:18.073406869 +0000 UTC m=+1184.353282785" Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.286772 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-774fb64f68-qgzkk"] Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.473739 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.586161 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-dns-svc\") pod \"7ca5eb4d-d2b1-4357-8b49-001763e93804\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.586332 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x9h4\" (UniqueName: \"kubernetes.io/projected/7ca5eb4d-d2b1-4357-8b49-001763e93804-kube-api-access-4x9h4\") pod \"7ca5eb4d-d2b1-4357-8b49-001763e93804\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.586451 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-ovsdbserver-sb\") pod \"7ca5eb4d-d2b1-4357-8b49-001763e93804\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.586888 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-dns-swift-storage-0\") pod \"7ca5eb4d-d2b1-4357-8b49-001763e93804\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.586946 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-config\") pod \"7ca5eb4d-d2b1-4357-8b49-001763e93804\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.587027 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-ovsdbserver-nb\") pod \"7ca5eb4d-d2b1-4357-8b49-001763e93804\" (UID: \"7ca5eb4d-d2b1-4357-8b49-001763e93804\") " Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.591198 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ca5eb4d-d2b1-4357-8b49-001763e93804-kube-api-access-4x9h4" (OuterVolumeSpecName: "kube-api-access-4x9h4") pod "7ca5eb4d-d2b1-4357-8b49-001763e93804" (UID: "7ca5eb4d-d2b1-4357-8b49-001763e93804"). InnerVolumeSpecName "kube-api-access-4x9h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.636611 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7ca5eb4d-d2b1-4357-8b49-001763e93804" (UID: "7ca5eb4d-d2b1-4357-8b49-001763e93804"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.647333 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7ca5eb4d-d2b1-4357-8b49-001763e93804" (UID: "7ca5eb4d-d2b1-4357-8b49-001763e93804"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.660944 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7ca5eb4d-d2b1-4357-8b49-001763e93804" (UID: "7ca5eb4d-d2b1-4357-8b49-001763e93804"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.664008 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7ca5eb4d-d2b1-4357-8b49-001763e93804" (UID: "7ca5eb4d-d2b1-4357-8b49-001763e93804"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.674441 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-config" (OuterVolumeSpecName: "config") pod "7ca5eb4d-d2b1-4357-8b49-001763e93804" (UID: "7ca5eb4d-d2b1-4357-8b49-001763e93804"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.689507 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.689542 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.689552 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.689561 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.689570 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ca5eb4d-d2b1-4357-8b49-001763e93804-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:18 crc kubenswrapper[4744]: I1003 16:44:18.689578 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x9h4\" (UniqueName: \"kubernetes.io/projected/7ca5eb4d-d2b1-4357-8b49-001763e93804-kube-api-access-4x9h4\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.071249 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-774fb64f68-qgzkk" event={"ID":"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e","Type":"ContainerStarted","Data":"21267c6080ad78c3f3e6a902cd2749e7e7e2d6e9becadc5d02f184cb171f45ee"} Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.071528 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-774fb64f68-qgzkk" event={"ID":"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e","Type":"ContainerStarted","Data":"b780b0753a10fbca106300463bb944daaae9fc5bca366bf7c51c9a7ae4bf758e"} Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.071548 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-774fb64f68-qgzkk" event={"ID":"0d2b9c37-6d79-420d-bc4e-3d5888b5d61e","Type":"ContainerStarted","Data":"21f1f09a5c8dda25ae2c509d93f5b82abaca380572dac578248b9b9553ee89cd"} Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.072157 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.073365 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.073916 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-krczc" event={"ID":"a8054599-30d8-42e0-bf45-2a8758b10486","Type":"ContainerStarted","Data":"7aa00cc7e148d70ea0f891901929ff502cafa279aba0996289a3c2ae059e648a"} Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.081115 4744 generic.go:334] "Generic (PLEG): container finished" podID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerID="7eb661e5353163eb93dc19558c766a2c4d32ad30ebec5298fa08a7861f1a24d1" exitCode=0 Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.081202 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4","Type":"ContainerDied","Data":"7eb661e5353163eb93dc19558c766a2c4d32ad30ebec5298fa08a7861f1a24d1"} Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.091477 4744 generic.go:334] "Generic (PLEG): container finished" podID="7ca5eb4d-d2b1-4357-8b49-001763e93804" containerID="42af22cf2a27fc349e7615656c18eb71e7b78bf252206da91e54c7bcca85a59a" exitCode=0 Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.091554 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" event={"ID":"7ca5eb4d-d2b1-4357-8b49-001763e93804","Type":"ContainerDied","Data":"42af22cf2a27fc349e7615656c18eb71e7b78bf252206da91e54c7bcca85a59a"} Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.091579 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" event={"ID":"7ca5eb4d-d2b1-4357-8b49-001763e93804","Type":"ContainerDied","Data":"acc9300193f94ed5e5b824c1a800330440d35a65ad1448f4fb20df9d1fce861d"} Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.091594 4744 scope.go:117] "RemoveContainer" containerID="42af22cf2a27fc349e7615656c18eb71e7b78bf252206da91e54c7bcca85a59a" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.091709 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-ndwh8" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.147977 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-krczc" podStartSLOduration=5.246253604 podStartE2EDuration="48.147954868s" podCreationTimestamp="2025-10-03 16:43:31 +0000 UTC" firstStartedPulling="2025-10-03 16:43:34.714263635 +0000 UTC m=+1140.994139531" lastFinishedPulling="2025-10-03 16:44:17.615964899 +0000 UTC m=+1183.895840795" observedRunningTime="2025-10-03 16:44:19.145919492 +0000 UTC m=+1185.425795388" watchObservedRunningTime="2025-10-03 16:44:19.147954868 +0000 UTC m=+1185.427830764" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.152030 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-774fb64f68-qgzkk" podStartSLOduration=2.15202149 podStartE2EDuration="2.15202149s" podCreationTimestamp="2025-10-03 16:44:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:19.108471923 +0000 UTC m=+1185.388347819" watchObservedRunningTime="2025-10-03 16:44:19.15202149 +0000 UTC m=+1185.431897386" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.158692 4744 scope.go:117] "RemoveContainer" containerID="f371500e59388e92174a2cefe793cdff8c9dc984766822aa684c5ba02fa610c3" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.188239 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-ndwh8"] Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.196442 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-ndwh8"] Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.234818 4744 scope.go:117] "RemoveContainer" containerID="42af22cf2a27fc349e7615656c18eb71e7b78bf252206da91e54c7bcca85a59a" Oct 03 16:44:19 crc kubenswrapper[4744]: E1003 16:44:19.244417 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42af22cf2a27fc349e7615656c18eb71e7b78bf252206da91e54c7bcca85a59a\": container with ID starting with 42af22cf2a27fc349e7615656c18eb71e7b78bf252206da91e54c7bcca85a59a not found: ID does not exist" containerID="42af22cf2a27fc349e7615656c18eb71e7b78bf252206da91e54c7bcca85a59a" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.244461 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42af22cf2a27fc349e7615656c18eb71e7b78bf252206da91e54c7bcca85a59a"} err="failed to get container status \"42af22cf2a27fc349e7615656c18eb71e7b78bf252206da91e54c7bcca85a59a\": rpc error: code = NotFound desc = could not find container \"42af22cf2a27fc349e7615656c18eb71e7b78bf252206da91e54c7bcca85a59a\": container with ID starting with 42af22cf2a27fc349e7615656c18eb71e7b78bf252206da91e54c7bcca85a59a not found: ID does not exist" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.244488 4744 scope.go:117] "RemoveContainer" containerID="f371500e59388e92174a2cefe793cdff8c9dc984766822aa684c5ba02fa610c3" Oct 03 16:44:19 crc kubenswrapper[4744]: E1003 16:44:19.245154 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f371500e59388e92174a2cefe793cdff8c9dc984766822aa684c5ba02fa610c3\": container with ID starting with f371500e59388e92174a2cefe793cdff8c9dc984766822aa684c5ba02fa610c3 not found: ID does not exist" containerID="f371500e59388e92174a2cefe793cdff8c9dc984766822aa684c5ba02fa610c3" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.245201 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f371500e59388e92174a2cefe793cdff8c9dc984766822aa684c5ba02fa610c3"} err="failed to get container status \"f371500e59388e92174a2cefe793cdff8c9dc984766822aa684c5ba02fa610c3\": rpc error: code = NotFound desc = could not find container \"f371500e59388e92174a2cefe793cdff8c9dc984766822aa684c5ba02fa610c3\": container with ID starting with f371500e59388e92174a2cefe793cdff8c9dc984766822aa684c5ba02fa610c3 not found: ID does not exist" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.360940 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.504181 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-sg-core-conf-yaml\") pod \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.504241 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-combined-ca-bundle\") pod \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.504308 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-log-httpd\") pod \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.504359 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-run-httpd\") pod \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.504431 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-scripts\") pod \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.504716 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqc65\" (UniqueName: \"kubernetes.io/projected/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-kube-api-access-qqc65\") pod \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.504785 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-config-data\") pod \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\" (UID: \"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4\") " Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.505243 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" (UID: "b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.506228 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" (UID: "b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.513673 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-scripts" (OuterVolumeSpecName: "scripts") pod "b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" (UID: "b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.517043 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-kube-api-access-qqc65" (OuterVolumeSpecName: "kube-api-access-qqc65") pod "b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" (UID: "b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4"). InnerVolumeSpecName "kube-api-access-qqc65". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.540595 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" (UID: "b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.596483 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" (UID: "b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.608550 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.608609 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.608624 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.608634 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.608646 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.608657 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqc65\" (UniqueName: \"kubernetes.io/projected/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-kube-api-access-qqc65\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.624817 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-config-data" (OuterVolumeSpecName: "config-data") pod "b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" (UID: "b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:19 crc kubenswrapper[4744]: I1003 16:44:19.712122 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.034592 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6589b6d4db-rbbtp" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.116239 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4","Type":"ContainerDied","Data":"a514666c722adef5d24f29fe37576007ea0b02af93a2eb43e825f763ced6e1eb"} Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.116562 4744 scope.go:117] "RemoveContainer" containerID="0b97444af620bd92a44ee61d76bf19333ee8ac3674e16796c07f625342ded3b3" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.116290 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.162527 4744 scope.go:117] "RemoveContainer" containerID="e56c8cf23f7e569c360c631b099ce8e5cd11460568423c06b51f9baaa1b6bac5" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.174559 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.200852 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.203132 4744 scope.go:117] "RemoveContainer" containerID="7eb661e5353163eb93dc19558c766a2c4d32ad30ebec5298fa08a7861f1a24d1" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.231241 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:20 crc kubenswrapper[4744]: E1003 16:44:20.232071 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ca5eb4d-d2b1-4357-8b49-001763e93804" containerName="init" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.232094 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ca5eb4d-d2b1-4357-8b49-001763e93804" containerName="init" Oct 03 16:44:20 crc kubenswrapper[4744]: E1003 16:44:20.232116 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="sg-core" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.232123 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="sg-core" Oct 03 16:44:20 crc kubenswrapper[4744]: E1003 16:44:20.232161 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="ceilometer-notification-agent" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.232173 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="ceilometer-notification-agent" Oct 03 16:44:20 crc kubenswrapper[4744]: E1003 16:44:20.232206 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="proxy-httpd" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.232216 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="proxy-httpd" Oct 03 16:44:20 crc kubenswrapper[4744]: E1003 16:44:20.232244 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="ceilometer-central-agent" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.232252 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="ceilometer-central-agent" Oct 03 16:44:20 crc kubenswrapper[4744]: E1003 16:44:20.232274 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ca5eb4d-d2b1-4357-8b49-001763e93804" containerName="dnsmasq-dns" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.232287 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ca5eb4d-d2b1-4357-8b49-001763e93804" containerName="dnsmasq-dns" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.232787 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="sg-core" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.232805 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ca5eb4d-d2b1-4357-8b49-001763e93804" containerName="dnsmasq-dns" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.232828 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="ceilometer-notification-agent" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.232843 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="ceilometer-central-agent" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.232862 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" containerName="proxy-httpd" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.236043 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.239390 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.239577 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.250253 4744 scope.go:117] "RemoveContainer" containerID="c4f9d07a32023e6cae16d77eac1219e6d66fcd4e519e8880aed981dd3e712ebe" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.251686 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.329705 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98n8n\" (UniqueName: \"kubernetes.io/projected/f26bf136-8ce5-4187-8e27-21d73d45319e-kube-api-access-98n8n\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.329833 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-scripts\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.329889 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.329942 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-config-data\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.329975 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.329999 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f26bf136-8ce5-4187-8e27-21d73d45319e-run-httpd\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.330039 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f26bf136-8ce5-4187-8e27-21d73d45319e-log-httpd\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.431762 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-scripts\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.431855 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.431915 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-config-data\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.431953 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.431983 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f26bf136-8ce5-4187-8e27-21d73d45319e-run-httpd\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.432028 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f26bf136-8ce5-4187-8e27-21d73d45319e-log-httpd\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.432058 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98n8n\" (UniqueName: \"kubernetes.io/projected/f26bf136-8ce5-4187-8e27-21d73d45319e-kube-api-access-98n8n\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.433596 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f26bf136-8ce5-4187-8e27-21d73d45319e-run-httpd\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.436852 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f26bf136-8ce5-4187-8e27-21d73d45319e-log-httpd\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.437116 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-scripts\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.439986 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.440462 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.440604 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-config-data\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.449237 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98n8n\" (UniqueName: \"kubernetes.io/projected/f26bf136-8ce5-4187-8e27-21d73d45319e-kube-api-access-98n8n\") pod \"ceilometer-0\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.595195 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.913876 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ca5eb4d-d2b1-4357-8b49-001763e93804" path="/var/lib/kubelet/pods/7ca5eb4d-d2b1-4357-8b49-001763e93804/volumes" Oct 03 16:44:20 crc kubenswrapper[4744]: I1003 16:44:20.915165 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4" path="/var/lib/kubelet/pods/b0354a0a-4c4d-4a1c-bc31-1ae5c7548cd4/volumes" Oct 03 16:44:21 crc kubenswrapper[4744]: I1003 16:44:21.125462 4744 generic.go:334] "Generic (PLEG): container finished" podID="dc8def8d-6149-4797-b453-103ec15579f6" containerID="8c2b637e9fea89323ced9b6a7c16ff4a1499d858e2fb5fcb8ba4fb3b95f37182" exitCode=0 Oct 03 16:44:21 crc kubenswrapper[4744]: I1003 16:44:21.125547 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6k6b8" event={"ID":"dc8def8d-6149-4797-b453-103ec15579f6","Type":"ContainerDied","Data":"8c2b637e9fea89323ced9b6a7c16ff4a1499d858e2fb5fcb8ba4fb3b95f37182"} Oct 03 16:44:21 crc kubenswrapper[4744]: I1003 16:44:21.154584 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:21 crc kubenswrapper[4744]: W1003 16:44:21.172372 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf26bf136_8ce5_4187_8e27_21d73d45319e.slice/crio-ee5b5f0bf40d5829a8a4b5495e113bb73daf2e94887e70d3cd3a84b281a30935 WatchSource:0}: Error finding container ee5b5f0bf40d5829a8a4b5495e113bb73daf2e94887e70d3cd3a84b281a30935: Status 404 returned error can't find the container with id ee5b5f0bf40d5829a8a4b5495e113bb73daf2e94887e70d3cd3a84b281a30935 Oct 03 16:44:22 crc kubenswrapper[4744]: I1003 16:44:22.136340 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f26bf136-8ce5-4187-8e27-21d73d45319e","Type":"ContainerStarted","Data":"ed780aed8b09d2751d833e6ade94da0942fa01b3405d9f734f64467bf7ac93aa"} Oct 03 16:44:22 crc kubenswrapper[4744]: I1003 16:44:22.137022 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f26bf136-8ce5-4187-8e27-21d73d45319e","Type":"ContainerStarted","Data":"ee5b5f0bf40d5829a8a4b5495e113bb73daf2e94887e70d3cd3a84b281a30935"} Oct 03 16:44:22 crc kubenswrapper[4744]: I1003 16:44:22.578674 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6k6b8" Oct 03 16:44:22 crc kubenswrapper[4744]: I1003 16:44:22.678090 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc8def8d-6149-4797-b453-103ec15579f6-combined-ca-bundle\") pod \"dc8def8d-6149-4797-b453-103ec15579f6\" (UID: \"dc8def8d-6149-4797-b453-103ec15579f6\") " Oct 03 16:44:22 crc kubenswrapper[4744]: I1003 16:44:22.678402 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqlt9\" (UniqueName: \"kubernetes.io/projected/dc8def8d-6149-4797-b453-103ec15579f6-kube-api-access-wqlt9\") pod \"dc8def8d-6149-4797-b453-103ec15579f6\" (UID: \"dc8def8d-6149-4797-b453-103ec15579f6\") " Oct 03 16:44:22 crc kubenswrapper[4744]: I1003 16:44:22.679074 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dc8def8d-6149-4797-b453-103ec15579f6-db-sync-config-data\") pod \"dc8def8d-6149-4797-b453-103ec15579f6\" (UID: \"dc8def8d-6149-4797-b453-103ec15579f6\") " Oct 03 16:44:22 crc kubenswrapper[4744]: I1003 16:44:22.683622 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc8def8d-6149-4797-b453-103ec15579f6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "dc8def8d-6149-4797-b453-103ec15579f6" (UID: "dc8def8d-6149-4797-b453-103ec15579f6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:22 crc kubenswrapper[4744]: I1003 16:44:22.704807 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc8def8d-6149-4797-b453-103ec15579f6-kube-api-access-wqlt9" (OuterVolumeSpecName: "kube-api-access-wqlt9") pod "dc8def8d-6149-4797-b453-103ec15579f6" (UID: "dc8def8d-6149-4797-b453-103ec15579f6"). InnerVolumeSpecName "kube-api-access-wqlt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:22 crc kubenswrapper[4744]: I1003 16:44:22.710330 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc8def8d-6149-4797-b453-103ec15579f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc8def8d-6149-4797-b453-103ec15579f6" (UID: "dc8def8d-6149-4797-b453-103ec15579f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:22 crc kubenswrapper[4744]: I1003 16:44:22.780774 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqlt9\" (UniqueName: \"kubernetes.io/projected/dc8def8d-6149-4797-b453-103ec15579f6-kube-api-access-wqlt9\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:22 crc kubenswrapper[4744]: I1003 16:44:22.780809 4744 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dc8def8d-6149-4797-b453-103ec15579f6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:22 crc kubenswrapper[4744]: I1003 16:44:22.780819 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc8def8d-6149-4797-b453-103ec15579f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.151175 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6k6b8" event={"ID":"dc8def8d-6149-4797-b453-103ec15579f6","Type":"ContainerDied","Data":"13aba0b99f1560496ecb508e3b1c42067ec0c4d2dcdfe7b1e50937a5be6a2716"} Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.152481 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13aba0b99f1560496ecb508e3b1c42067ec0c4d2dcdfe7b1e50937a5be6a2716" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.151187 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6k6b8" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.157904 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f26bf136-8ce5-4187-8e27-21d73d45319e","Type":"ContainerStarted","Data":"9b3373ebc9cf64f92e3f7fcce4bbe5d1a5fb67bb5be2f8e1a20193f4c57caabd"} Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.369644 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr"] Oct 03 16:44:23 crc kubenswrapper[4744]: E1003 16:44:23.370023 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc8def8d-6149-4797-b453-103ec15579f6" containerName="barbican-db-sync" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.370035 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc8def8d-6149-4797-b453-103ec15579f6" containerName="barbican-db-sync" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.370192 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc8def8d-6149-4797-b453-103ec15579f6" containerName="barbican-db-sync" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.371014 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.373795 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.374090 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.379561 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-ttscr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.405874 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5c46748695-jn5bb"] Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.420830 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.447306 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr"] Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.448156 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.533941 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5c46748695-jn5bb"] Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.539925 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b4d5e00-5174-4dc3-b68e-185dedb764f1-config-data\") pod \"barbican-keystone-listener-6dfd6c4c58-rr7fr\" (UID: \"6b4d5e00-5174-4dc3-b68e-185dedb764f1\") " pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.540083 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b4d5e00-5174-4dc3-b68e-185dedb764f1-config-data-custom\") pod \"barbican-keystone-listener-6dfd6c4c58-rr7fr\" (UID: \"6b4d5e00-5174-4dc3-b68e-185dedb764f1\") " pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.540159 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b4d5e00-5174-4dc3-b68e-185dedb764f1-logs\") pod \"barbican-keystone-listener-6dfd6c4c58-rr7fr\" (UID: \"6b4d5e00-5174-4dc3-b68e-185dedb764f1\") " pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.540219 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntgbm\" (UniqueName: \"kubernetes.io/projected/9c69aee3-5a63-4011-b4ff-0afca3c16880-kube-api-access-ntgbm\") pod \"barbican-worker-5c46748695-jn5bb\" (UID: \"9c69aee3-5a63-4011-b4ff-0afca3c16880\") " pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.540290 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhr2p\" (UniqueName: \"kubernetes.io/projected/6b4d5e00-5174-4dc3-b68e-185dedb764f1-kube-api-access-xhr2p\") pod \"barbican-keystone-listener-6dfd6c4c58-rr7fr\" (UID: \"6b4d5e00-5174-4dc3-b68e-185dedb764f1\") " pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.540390 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b4d5e00-5174-4dc3-b68e-185dedb764f1-combined-ca-bundle\") pod \"barbican-keystone-listener-6dfd6c4c58-rr7fr\" (UID: \"6b4d5e00-5174-4dc3-b68e-185dedb764f1\") " pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.540413 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c69aee3-5a63-4011-b4ff-0afca3c16880-logs\") pod \"barbican-worker-5c46748695-jn5bb\" (UID: \"9c69aee3-5a63-4011-b4ff-0afca3c16880\") " pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.540436 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c69aee3-5a63-4011-b4ff-0afca3c16880-combined-ca-bundle\") pod \"barbican-worker-5c46748695-jn5bb\" (UID: \"9c69aee3-5a63-4011-b4ff-0afca3c16880\") " pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.540457 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c69aee3-5a63-4011-b4ff-0afca3c16880-config-data\") pod \"barbican-worker-5c46748695-jn5bb\" (UID: \"9c69aee3-5a63-4011-b4ff-0afca3c16880\") " pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.540541 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c69aee3-5a63-4011-b4ff-0afca3c16880-config-data-custom\") pod \"barbican-worker-5c46748695-jn5bb\" (UID: \"9c69aee3-5a63-4011-b4ff-0afca3c16880\") " pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.637342 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-nnv9h"] Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.639136 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641443 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-config\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641481 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641519 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b4d5e00-5174-4dc3-b68e-185dedb764f1-logs\") pod \"barbican-keystone-listener-6dfd6c4c58-rr7fr\" (UID: \"6b4d5e00-5174-4dc3-b68e-185dedb764f1\") " pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641550 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntgbm\" (UniqueName: \"kubernetes.io/projected/9c69aee3-5a63-4011-b4ff-0afca3c16880-kube-api-access-ntgbm\") pod \"barbican-worker-5c46748695-jn5bb\" (UID: \"9c69aee3-5a63-4011-b4ff-0afca3c16880\") " pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641582 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhr2p\" (UniqueName: \"kubernetes.io/projected/6b4d5e00-5174-4dc3-b68e-185dedb764f1-kube-api-access-xhr2p\") pod \"barbican-keystone-listener-6dfd6c4c58-rr7fr\" (UID: \"6b4d5e00-5174-4dc3-b68e-185dedb764f1\") " pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641601 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n52c\" (UniqueName: \"kubernetes.io/projected/862d1595-46ba-4e56-8b99-2ede960d343c-kube-api-access-5n52c\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641620 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641637 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641673 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b4d5e00-5174-4dc3-b68e-185dedb764f1-combined-ca-bundle\") pod \"barbican-keystone-listener-6dfd6c4c58-rr7fr\" (UID: \"6b4d5e00-5174-4dc3-b68e-185dedb764f1\") " pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641688 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c69aee3-5a63-4011-b4ff-0afca3c16880-logs\") pod \"barbican-worker-5c46748695-jn5bb\" (UID: \"9c69aee3-5a63-4011-b4ff-0afca3c16880\") " pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641707 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c69aee3-5a63-4011-b4ff-0afca3c16880-combined-ca-bundle\") pod \"barbican-worker-5c46748695-jn5bb\" (UID: \"9c69aee3-5a63-4011-b4ff-0afca3c16880\") " pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641721 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c69aee3-5a63-4011-b4ff-0afca3c16880-config-data\") pod \"barbican-worker-5c46748695-jn5bb\" (UID: \"9c69aee3-5a63-4011-b4ff-0afca3c16880\") " pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641754 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c69aee3-5a63-4011-b4ff-0afca3c16880-config-data-custom\") pod \"barbican-worker-5c46748695-jn5bb\" (UID: \"9c69aee3-5a63-4011-b4ff-0afca3c16880\") " pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641780 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641796 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b4d5e00-5174-4dc3-b68e-185dedb764f1-config-data\") pod \"barbican-keystone-listener-6dfd6c4c58-rr7fr\" (UID: \"6b4d5e00-5174-4dc3-b68e-185dedb764f1\") " pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.641832 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b4d5e00-5174-4dc3-b68e-185dedb764f1-config-data-custom\") pod \"barbican-keystone-listener-6dfd6c4c58-rr7fr\" (UID: \"6b4d5e00-5174-4dc3-b68e-185dedb764f1\") " pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.642810 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b4d5e00-5174-4dc3-b68e-185dedb764f1-logs\") pod \"barbican-keystone-listener-6dfd6c4c58-rr7fr\" (UID: \"6b4d5e00-5174-4dc3-b68e-185dedb764f1\") " pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.644868 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c69aee3-5a63-4011-b4ff-0afca3c16880-logs\") pod \"barbican-worker-5c46748695-jn5bb\" (UID: \"9c69aee3-5a63-4011-b4ff-0afca3c16880\") " pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.646039 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b4d5e00-5174-4dc3-b68e-185dedb764f1-config-data-custom\") pod \"barbican-keystone-listener-6dfd6c4c58-rr7fr\" (UID: \"6b4d5e00-5174-4dc3-b68e-185dedb764f1\") " pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.658910 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b4d5e00-5174-4dc3-b68e-185dedb764f1-combined-ca-bundle\") pod \"barbican-keystone-listener-6dfd6c4c58-rr7fr\" (UID: \"6b4d5e00-5174-4dc3-b68e-185dedb764f1\") " pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.672836 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c69aee3-5a63-4011-b4ff-0afca3c16880-config-data\") pod \"barbican-worker-5c46748695-jn5bb\" (UID: \"9c69aee3-5a63-4011-b4ff-0afca3c16880\") " pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.673803 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-nnv9h"] Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.675370 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhr2p\" (UniqueName: \"kubernetes.io/projected/6b4d5e00-5174-4dc3-b68e-185dedb764f1-kube-api-access-xhr2p\") pod \"barbican-keystone-listener-6dfd6c4c58-rr7fr\" (UID: \"6b4d5e00-5174-4dc3-b68e-185dedb764f1\") " pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.678192 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c69aee3-5a63-4011-b4ff-0afca3c16880-combined-ca-bundle\") pod \"barbican-worker-5c46748695-jn5bb\" (UID: \"9c69aee3-5a63-4011-b4ff-0afca3c16880\") " pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.678759 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c69aee3-5a63-4011-b4ff-0afca3c16880-config-data-custom\") pod \"barbican-worker-5c46748695-jn5bb\" (UID: \"9c69aee3-5a63-4011-b4ff-0afca3c16880\") " pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.680444 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntgbm\" (UniqueName: \"kubernetes.io/projected/9c69aee3-5a63-4011-b4ff-0afca3c16880-kube-api-access-ntgbm\") pod \"barbican-worker-5c46748695-jn5bb\" (UID: \"9c69aee3-5a63-4011-b4ff-0afca3c16880\") " pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.687416 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b4d5e00-5174-4dc3-b68e-185dedb764f1-config-data\") pod \"barbican-keystone-listener-6dfd6c4c58-rr7fr\" (UID: \"6b4d5e00-5174-4dc3-b68e-185dedb764f1\") " pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.696137 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.698299 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5467484d76-wdx4m"] Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.700485 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.702988 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.716711 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5467484d76-wdx4m"] Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.747564 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n52c\" (UniqueName: \"kubernetes.io/projected/862d1595-46ba-4e56-8b99-2ede960d343c-kube-api-access-5n52c\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.747634 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.747852 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.747883 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-config-data-custom\") pod \"barbican-api-5467484d76-wdx4m\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.747919 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-combined-ca-bundle\") pod \"barbican-api-5467484d76-wdx4m\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.747948 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-config-data\") pod \"barbican-api-5467484d76-wdx4m\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.748013 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-logs\") pod \"barbican-api-5467484d76-wdx4m\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.748045 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.748203 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-config\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.748249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.748283 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b2ks\" (UniqueName: \"kubernetes.io/projected/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-kube-api-access-6b2ks\") pod \"barbican-api-5467484d76-wdx4m\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.752343 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.752879 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.753202 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.753607 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-config\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.753727 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.757959 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c46748695-jn5bb" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.824939 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n52c\" (UniqueName: \"kubernetes.io/projected/862d1595-46ba-4e56-8b99-2ede960d343c-kube-api-access-5n52c\") pod \"dnsmasq-dns-75c8ddd69c-nnv9h\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.859569 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b2ks\" (UniqueName: \"kubernetes.io/projected/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-kube-api-access-6b2ks\") pod \"barbican-api-5467484d76-wdx4m\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.860027 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-config-data-custom\") pod \"barbican-api-5467484d76-wdx4m\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.860085 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-combined-ca-bundle\") pod \"barbican-api-5467484d76-wdx4m\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.860128 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-config-data\") pod \"barbican-api-5467484d76-wdx4m\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.860388 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-logs\") pod \"barbican-api-5467484d76-wdx4m\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.867591 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-logs\") pod \"barbican-api-5467484d76-wdx4m\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.868034 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-config-data-custom\") pod \"barbican-api-5467484d76-wdx4m\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.875188 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-combined-ca-bundle\") pod \"barbican-api-5467484d76-wdx4m\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.887703 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b2ks\" (UniqueName: \"kubernetes.io/projected/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-kube-api-access-6b2ks\") pod \"barbican-api-5467484d76-wdx4m\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:23 crc kubenswrapper[4744]: I1003 16:44:23.891987 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-config-data\") pod \"barbican-api-5467484d76-wdx4m\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:24 crc kubenswrapper[4744]: I1003 16:44:24.115135 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:24 crc kubenswrapper[4744]: I1003 16:44:24.123120 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:24 crc kubenswrapper[4744]: I1003 16:44:24.213693 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f26bf136-8ce5-4187-8e27-21d73d45319e","Type":"ContainerStarted","Data":"4c28569118247e74e8bffc902ce85307d90b95985f47f1c1317d76e122f0200b"} Oct 03 16:44:24 crc kubenswrapper[4744]: I1003 16:44:24.257371 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5c46748695-jn5bb"] Oct 03 16:44:24 crc kubenswrapper[4744]: I1003 16:44:24.274306 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr"] Oct 03 16:44:24 crc kubenswrapper[4744]: W1003 16:44:24.287596 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c69aee3_5a63_4011_b4ff_0afca3c16880.slice/crio-2ef0919978719034e313249614025f78789a3ff50effe161a66a2ae11fb70a7f WatchSource:0}: Error finding container 2ef0919978719034e313249614025f78789a3ff50effe161a66a2ae11fb70a7f: Status 404 returned error can't find the container with id 2ef0919978719034e313249614025f78789a3ff50effe161a66a2ae11fb70a7f Oct 03 16:44:24 crc kubenswrapper[4744]: I1003 16:44:24.749982 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5467484d76-wdx4m"] Oct 03 16:44:24 crc kubenswrapper[4744]: I1003 16:44:24.759356 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-nnv9h"] Oct 03 16:44:24 crc kubenswrapper[4744]: I1003 16:44:24.998004 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 03 16:44:24 crc kubenswrapper[4744]: I1003 16:44:24.999481 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.008270 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.008431 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.008554 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-lrl8f" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.027762 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.111058 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1-openstack-config-secret\") pod \"openstackclient\" (UID: \"38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1\") " pod="openstack/openstackclient" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.111263 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1\") " pod="openstack/openstackclient" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.111375 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1-openstack-config\") pod \"openstackclient\" (UID: \"38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1\") " pod="openstack/openstackclient" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.111531 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrx97\" (UniqueName: \"kubernetes.io/projected/38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1-kube-api-access-rrx97\") pod \"openstackclient\" (UID: \"38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1\") " pod="openstack/openstackclient" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.213006 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1\") " pod="openstack/openstackclient" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.213106 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1-openstack-config\") pod \"openstackclient\" (UID: \"38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1\") " pod="openstack/openstackclient" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.213626 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrx97\" (UniqueName: \"kubernetes.io/projected/38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1-kube-api-access-rrx97\") pod \"openstackclient\" (UID: \"38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1\") " pod="openstack/openstackclient" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.213706 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1-openstack-config-secret\") pod \"openstackclient\" (UID: \"38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1\") " pod="openstack/openstackclient" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.222916 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1-openstack-config\") pod \"openstackclient\" (UID: \"38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1\") " pod="openstack/openstackclient" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.224522 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1-openstack-config-secret\") pod \"openstackclient\" (UID: \"38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1\") " pod="openstack/openstackclient" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.229395 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1\") " pod="openstack/openstackclient" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.236404 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrx97\" (UniqueName: \"kubernetes.io/projected/38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1-kube-api-access-rrx97\") pod \"openstackclient\" (UID: \"38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1\") " pod="openstack/openstackclient" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.246657 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c46748695-jn5bb" event={"ID":"9c69aee3-5a63-4011-b4ff-0afca3c16880","Type":"ContainerStarted","Data":"2ef0919978719034e313249614025f78789a3ff50effe161a66a2ae11fb70a7f"} Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.249693 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5467484d76-wdx4m" event={"ID":"102bbbfd-8dd8-4946-9f35-a22a87ebaaed","Type":"ContainerStarted","Data":"8a8ff0666730ce0e751b72083f365fc1952aa3b0d6aeba77566dba6aead408c9"} Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.249777 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5467484d76-wdx4m" event={"ID":"102bbbfd-8dd8-4946-9f35-a22a87ebaaed","Type":"ContainerStarted","Data":"77d1c3e3a61cfe2d3ed30b7fb39f2af9e7f613712d7f3921d9fa9e1efdc7819c"} Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.265328 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" event={"ID":"862d1595-46ba-4e56-8b99-2ede960d343c","Type":"ContainerStarted","Data":"2a3d0286bf2ff14c339d2d4b6446b3b9d667099252294fdf22a10fd35cf2fb60"} Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.265385 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" event={"ID":"862d1595-46ba-4e56-8b99-2ede960d343c","Type":"ContainerStarted","Data":"e9ea27c666084ce7b262c680ac76fe3bd2269bd25fba2e8175d3401135530cad"} Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.269639 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" event={"ID":"6b4d5e00-5174-4dc3-b68e-185dedb764f1","Type":"ContainerStarted","Data":"37cd23a9517a0e03b59af8e242ef0e5d1e07ccfffd18b952aae12e17afdcab09"} Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.332788 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 16:44:25 crc kubenswrapper[4744]: I1003 16:44:25.964374 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.298160 4744 generic.go:334] "Generic (PLEG): container finished" podID="862d1595-46ba-4e56-8b99-2ede960d343c" containerID="2a3d0286bf2ff14c339d2d4b6446b3b9d667099252294fdf22a10fd35cf2fb60" exitCode=0 Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.298620 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" event={"ID":"862d1595-46ba-4e56-8b99-2ede960d343c","Type":"ContainerDied","Data":"2a3d0286bf2ff14c339d2d4b6446b3b9d667099252294fdf22a10fd35cf2fb60"} Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.303334 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1","Type":"ContainerStarted","Data":"09a9990469e34199568b291353aca63df1fac3bf6b2372a7a99ba53cf33997f8"} Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.310939 4744 generic.go:334] "Generic (PLEG): container finished" podID="a8054599-30d8-42e0-bf45-2a8758b10486" containerID="7aa00cc7e148d70ea0f891901929ff502cafa279aba0996289a3c2ae059e648a" exitCode=0 Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.310985 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-krczc" event={"ID":"a8054599-30d8-42e0-bf45-2a8758b10486","Type":"ContainerDied","Data":"7aa00cc7e148d70ea0f891901929ff502cafa279aba0996289a3c2ae059e648a"} Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.319740 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5467484d76-wdx4m" event={"ID":"102bbbfd-8dd8-4946-9f35-a22a87ebaaed","Type":"ContainerStarted","Data":"96bf8b18b0a9bf7861ee3471bb012ef363c641506d0431562a761c26266e3de2"} Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.320653 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.320718 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.370459 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5467484d76-wdx4m" podStartSLOduration=3.370438043 podStartE2EDuration="3.370438043s" podCreationTimestamp="2025-10-03 16:44:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:26.359473975 +0000 UTC m=+1192.639349881" watchObservedRunningTime="2025-10-03 16:44:26.370438043 +0000 UTC m=+1192.650313939" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.687531 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7bf8c46bb6-7m9v6"] Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.689270 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.691067 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.691886 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.713050 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7bf8c46bb6-7m9v6"] Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.747124 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/068cb298-fa9c-4e22-92cc-0aa31310f185-config-data-custom\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.747470 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/068cb298-fa9c-4e22-92cc-0aa31310f185-internal-tls-certs\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.747703 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/068cb298-fa9c-4e22-92cc-0aa31310f185-logs\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.747812 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068cb298-fa9c-4e22-92cc-0aa31310f185-config-data\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.747920 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/068cb298-fa9c-4e22-92cc-0aa31310f185-public-tls-certs\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.748081 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kggj\" (UniqueName: \"kubernetes.io/projected/068cb298-fa9c-4e22-92cc-0aa31310f185-kube-api-access-5kggj\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.748198 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068cb298-fa9c-4e22-92cc-0aa31310f185-combined-ca-bundle\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.854540 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068cb298-fa9c-4e22-92cc-0aa31310f185-combined-ca-bundle\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.854617 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/068cb298-fa9c-4e22-92cc-0aa31310f185-config-data-custom\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.854844 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/068cb298-fa9c-4e22-92cc-0aa31310f185-internal-tls-certs\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.854877 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/068cb298-fa9c-4e22-92cc-0aa31310f185-logs\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.854898 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068cb298-fa9c-4e22-92cc-0aa31310f185-config-data\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.854914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/068cb298-fa9c-4e22-92cc-0aa31310f185-public-tls-certs\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.854942 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kggj\" (UniqueName: \"kubernetes.io/projected/068cb298-fa9c-4e22-92cc-0aa31310f185-kube-api-access-5kggj\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.860084 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/068cb298-fa9c-4e22-92cc-0aa31310f185-combined-ca-bundle\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.868336 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/068cb298-fa9c-4e22-92cc-0aa31310f185-config-data-custom\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.868694 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/068cb298-fa9c-4e22-92cc-0aa31310f185-logs\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.868927 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/068cb298-fa9c-4e22-92cc-0aa31310f185-internal-tls-certs\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.877748 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kggj\" (UniqueName: \"kubernetes.io/projected/068cb298-fa9c-4e22-92cc-0aa31310f185-kube-api-access-5kggj\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.880647 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/068cb298-fa9c-4e22-92cc-0aa31310f185-config-data\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:26 crc kubenswrapper[4744]: I1003 16:44:26.882134 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/068cb298-fa9c-4e22-92cc-0aa31310f185-public-tls-certs\") pod \"barbican-api-7bf8c46bb6-7m9v6\" (UID: \"068cb298-fa9c-4e22-92cc-0aa31310f185\") " pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.032575 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.343850 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c46748695-jn5bb" event={"ID":"9c69aee3-5a63-4011-b4ff-0afca3c16880","Type":"ContainerStarted","Data":"e3fd66065c4666b46c9913354eee079e655bfa10ef6a2102e0607640b79b93de"} Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.354950 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.377818 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" event={"ID":"6b4d5e00-5174-4dc3-b68e-185dedb764f1","Type":"ContainerStarted","Data":"934f6682d1f98f03f5942008aa4a330fec4aa9c181cd839bd0458961c8f0d839"} Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.384797 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" podStartSLOduration=4.384781526 podStartE2EDuration="4.384781526s" podCreationTimestamp="2025-10-03 16:44:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:27.378132576 +0000 UTC m=+1193.658008482" watchObservedRunningTime="2025-10-03 16:44:27.384781526 +0000 UTC m=+1193.664657422" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.653394 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7bf8c46bb6-7m9v6"] Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.710116 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-krczc" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.787423 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-scripts\") pod \"a8054599-30d8-42e0-bf45-2a8758b10486\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.787509 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99bm5\" (UniqueName: \"kubernetes.io/projected/a8054599-30d8-42e0-bf45-2a8758b10486-kube-api-access-99bm5\") pod \"a8054599-30d8-42e0-bf45-2a8758b10486\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.787555 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-db-sync-config-data\") pod \"a8054599-30d8-42e0-bf45-2a8758b10486\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.787599 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-config-data\") pod \"a8054599-30d8-42e0-bf45-2a8758b10486\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.787694 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8054599-30d8-42e0-bf45-2a8758b10486-etc-machine-id\") pod \"a8054599-30d8-42e0-bf45-2a8758b10486\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.787798 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-combined-ca-bundle\") pod \"a8054599-30d8-42e0-bf45-2a8758b10486\" (UID: \"a8054599-30d8-42e0-bf45-2a8758b10486\") " Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.788926 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8054599-30d8-42e0-bf45-2a8758b10486-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a8054599-30d8-42e0-bf45-2a8758b10486" (UID: "a8054599-30d8-42e0-bf45-2a8758b10486"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.798704 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a8054599-30d8-42e0-bf45-2a8758b10486" (UID: "a8054599-30d8-42e0-bf45-2a8758b10486"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.798675 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8054599-30d8-42e0-bf45-2a8758b10486-kube-api-access-99bm5" (OuterVolumeSpecName: "kube-api-access-99bm5") pod "a8054599-30d8-42e0-bf45-2a8758b10486" (UID: "a8054599-30d8-42e0-bf45-2a8758b10486"). InnerVolumeSpecName "kube-api-access-99bm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.825804 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-scripts" (OuterVolumeSpecName: "scripts") pod "a8054599-30d8-42e0-bf45-2a8758b10486" (UID: "a8054599-30d8-42e0-bf45-2a8758b10486"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.844677 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8054599-30d8-42e0-bf45-2a8758b10486" (UID: "a8054599-30d8-42e0-bf45-2a8758b10486"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.864595 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-config-data" (OuterVolumeSpecName: "config-data") pod "a8054599-30d8-42e0-bf45-2a8758b10486" (UID: "a8054599-30d8-42e0-bf45-2a8758b10486"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.892787 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.892814 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.892823 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99bm5\" (UniqueName: \"kubernetes.io/projected/a8054599-30d8-42e0-bf45-2a8758b10486-kube-api-access-99bm5\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.892832 4744 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.892840 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8054599-30d8-42e0-bf45-2a8758b10486-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:27 crc kubenswrapper[4744]: I1003 16:44:27.892856 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8054599-30d8-42e0-bf45-2a8758b10486-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.238447 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-689c67db59-x766r"] Oct 03 16:44:28 crc kubenswrapper[4744]: E1003 16:44:28.238900 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8054599-30d8-42e0-bf45-2a8758b10486" containerName="cinder-db-sync" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.238917 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8054599-30d8-42e0-bf45-2a8758b10486" containerName="cinder-db-sync" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.239090 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8054599-30d8-42e0-bf45-2a8758b10486" containerName="cinder-db-sync" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.240107 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.256169 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-689c67db59-x766r"] Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.262734 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.262954 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.263208 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.300413 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c64dd03-9cb9-402f-9697-a23538c9ebc5-log-httpd\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.300514 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c64dd03-9cb9-402f-9697-a23538c9ebc5-run-httpd\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.300559 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1c64dd03-9cb9-402f-9697-a23538c9ebc5-etc-swift\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.300578 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c64dd03-9cb9-402f-9697-a23538c9ebc5-config-data\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.300597 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c64dd03-9cb9-402f-9697-a23538c9ebc5-public-tls-certs\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.300617 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c64dd03-9cb9-402f-9697-a23538c9ebc5-combined-ca-bundle\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.300641 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5qc7\" (UniqueName: \"kubernetes.io/projected/1c64dd03-9cb9-402f-9697-a23538c9ebc5-kube-api-access-n5qc7\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.300672 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c64dd03-9cb9-402f-9697-a23538c9ebc5-internal-tls-certs\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.391202 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f26bf136-8ce5-4187-8e27-21d73d45319e","Type":"ContainerStarted","Data":"df955e1affcb1641a0c07eb8a16a2befa101d98292eafdb0f18b034900de85c5"} Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.391323 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.393173 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c46748695-jn5bb" event={"ID":"9c69aee3-5a63-4011-b4ff-0afca3c16880","Type":"ContainerStarted","Data":"9eb8b08cb031029c9d8dcc0bd5a454fb64b9cc6048c651c95f78d29e263d4a73"} Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.396338 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-krczc" event={"ID":"a8054599-30d8-42e0-bf45-2a8758b10486","Type":"ContainerDied","Data":"cf120b7b85db73f9b55659238209e60a464ef5438c7aa01a1a2e471e28c33d7d"} Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.396363 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf120b7b85db73f9b55659238209e60a464ef5438c7aa01a1a2e471e28c33d7d" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.396433 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-krczc" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.406734 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1c64dd03-9cb9-402f-9697-a23538c9ebc5-etc-swift\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.406767 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c64dd03-9cb9-402f-9697-a23538c9ebc5-config-data\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.406791 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c64dd03-9cb9-402f-9697-a23538c9ebc5-public-tls-certs\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.406818 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c64dd03-9cb9-402f-9697-a23538c9ebc5-combined-ca-bundle\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.406847 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5qc7\" (UniqueName: \"kubernetes.io/projected/1c64dd03-9cb9-402f-9697-a23538c9ebc5-kube-api-access-n5qc7\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.406904 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c64dd03-9cb9-402f-9697-a23538c9ebc5-internal-tls-certs\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.406994 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c64dd03-9cb9-402f-9697-a23538c9ebc5-log-httpd\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.407032 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c64dd03-9cb9-402f-9697-a23538c9ebc5-run-httpd\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.407479 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c64dd03-9cb9-402f-9697-a23538c9ebc5-run-httpd\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.416930 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1c64dd03-9cb9-402f-9697-a23538c9ebc5-etc-swift\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.423442 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" event={"ID":"862d1595-46ba-4e56-8b99-2ede960d343c","Type":"ContainerStarted","Data":"a1984dd5069bb5139f99baf81a8d0a6eb29a1d77b9401ddf6b28286dcd983fcc"} Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.426416 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c64dd03-9cb9-402f-9697-a23538c9ebc5-combined-ca-bundle\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.426804 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c64dd03-9cb9-402f-9697-a23538c9ebc5-config-data\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.428669 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c64dd03-9cb9-402f-9697-a23538c9ebc5-log-httpd\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.433030 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c64dd03-9cb9-402f-9697-a23538c9ebc5-public-tls-certs\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.433675 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" event={"ID":"6b4d5e00-5174-4dc3-b68e-185dedb764f1","Type":"ContainerStarted","Data":"25c3956a6d68f62b9ec05cfb6a5c0899f7bae483e7d350c5cd916b3261cdf686"} Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.434200 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c64dd03-9cb9-402f-9697-a23538c9ebc5-internal-tls-certs\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.449180 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5qc7\" (UniqueName: \"kubernetes.io/projected/1c64dd03-9cb9-402f-9697-a23538c9ebc5-kube-api-access-n5qc7\") pod \"swift-proxy-689c67db59-x766r\" (UID: \"1c64dd03-9cb9-402f-9697-a23538c9ebc5\") " pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.459444 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bf8c46bb6-7m9v6" event={"ID":"068cb298-fa9c-4e22-92cc-0aa31310f185","Type":"ContainerStarted","Data":"65ed2aa4b5d073a839f8cfb8ca799e3c180da78751e671d59b5bbd1612860af7"} Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.459487 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bf8c46bb6-7m9v6" event={"ID":"068cb298-fa9c-4e22-92cc-0aa31310f185","Type":"ContainerStarted","Data":"10423f6e41b1af844da64c3f8d8fccd05faa449321ae65e007d4704fcfa740cf"} Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.459550 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bf8c46bb6-7m9v6" event={"ID":"068cb298-fa9c-4e22-92cc-0aa31310f185","Type":"ContainerStarted","Data":"cd46caae120824626a5e7a0f83bf352bd84646b282514a6854fd0a4d96c3df59"} Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.460621 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.460645 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.469416 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.812154701 podStartE2EDuration="8.469394274s" podCreationTimestamp="2025-10-03 16:44:20 +0000 UTC" firstStartedPulling="2025-10-03 16:44:21.175331846 +0000 UTC m=+1187.455207752" lastFinishedPulling="2025-10-03 16:44:26.832571429 +0000 UTC m=+1193.112447325" observedRunningTime="2025-10-03 16:44:28.433819977 +0000 UTC m=+1194.713695873" watchObservedRunningTime="2025-10-03 16:44:28.469394274 +0000 UTC m=+1194.749270170" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.476863 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5c46748695-jn5bb" podStartSLOduration=2.941415247 podStartE2EDuration="5.476845463s" podCreationTimestamp="2025-10-03 16:44:23 +0000 UTC" firstStartedPulling="2025-10-03 16:44:24.297793788 +0000 UTC m=+1190.577669704" lastFinishedPulling="2025-10-03 16:44:26.833224024 +0000 UTC m=+1193.113099920" observedRunningTime="2025-10-03 16:44:28.472425942 +0000 UTC m=+1194.752301838" watchObservedRunningTime="2025-10-03 16:44:28.476845463 +0000 UTC m=+1194.756721359" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.513067 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6dfd6c4c58-rr7fr" podStartSLOduration=2.9628564219999998 podStartE2EDuration="5.513053183s" podCreationTimestamp="2025-10-03 16:44:23 +0000 UTC" firstStartedPulling="2025-10-03 16:44:24.287630468 +0000 UTC m=+1190.567506354" lastFinishedPulling="2025-10-03 16:44:26.837827219 +0000 UTC m=+1193.117703115" observedRunningTime="2025-10-03 16:44:28.51288181 +0000 UTC m=+1194.792757706" watchObservedRunningTime="2025-10-03 16:44:28.513053183 +0000 UTC m=+1194.792929079" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.559839 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.563338 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7bf8c46bb6-7m9v6" podStartSLOduration=2.563317893 podStartE2EDuration="2.563317893s" podCreationTimestamp="2025-10-03 16:44:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:28.550715777 +0000 UTC m=+1194.830591673" watchObservedRunningTime="2025-10-03 16:44:28.563317893 +0000 UTC m=+1194.843193789" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.632483 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.634108 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.640969 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.641240 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8r7z7" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.643054 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.644449 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.651436 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.683826 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-nnv9h"] Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.715133 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qcrs4"] Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.716832 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.717715 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgpfl\" (UniqueName: \"kubernetes.io/projected/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-kube-api-access-mgpfl\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.717781 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-scripts\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.717799 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.717832 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.717856 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-config-data\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.717909 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.763595 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qcrs4"] Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.775942 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.777995 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.791746 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.793404 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.793766 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.802985 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.819418 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.819683 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.819764 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-scripts\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.819861 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-sys\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.819931 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-dev\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.819997 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.820076 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-sys\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.820149 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgpfl\" (UniqueName: \"kubernetes.io/projected/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-kube-api-access-mgpfl\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.820230 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.820297 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.820379 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.820447 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mpmv\" (UniqueName: \"kubernetes.io/projected/fd08b448-b737-4aa8-b263-b48e3c32f03b-kube-api-access-2mpmv\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.820550 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-dev\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.820697 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.820767 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/65ff9f68-860d-41e6-a48b-4ada29ea6275-ceph\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.820842 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-scripts\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.820908 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.820974 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.821049 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-config-data-custom\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.821117 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-config\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.821179 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-run\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.821260 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-nvme\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.821338 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.821406 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.821471 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.821560 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.821636 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-run\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.821708 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-config-data\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.821780 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-config-data\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.821854 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-lib-modules\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.821921 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.821989 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.822070 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.822143 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/fd08b448-b737-4aa8-b263-b48e3c32f03b-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.822207 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.822274 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-dns-svc\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.822343 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.822416 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.822508 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.822655 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppjtf\" (UniqueName: \"kubernetes.io/projected/4047d3e2-47ab-412e-a491-34835eac6716-kube-api-access-ppjtf\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.822723 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.822791 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.822875 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.822950 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj4sb\" (UniqueName: \"kubernetes.io/projected/65ff9f68-860d-41e6-a48b-4ada29ea6275-kube-api-access-lj4sb\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.828829 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.839182 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-scripts\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.843522 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.844929 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-config-data\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.861227 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.881058 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.885113 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgpfl\" (UniqueName: \"kubernetes.io/projected/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-kube-api-access-mgpfl\") pod \"cinder-scheduler-0\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.925566 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-dev\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.925801 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.925918 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-sys\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.926023 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.926107 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.931702 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.931885 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mpmv\" (UniqueName: \"kubernetes.io/projected/fd08b448-b737-4aa8-b263-b48e3c32f03b-kube-api-access-2mpmv\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.932090 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-dev\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.932211 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/65ff9f68-860d-41e6-a48b-4ada29ea6275-ceph\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.932325 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.932446 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.932608 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-config-data-custom\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.936827 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-config\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.936920 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-run\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.926274 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-dev\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.927935 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.938065 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-config\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.935249 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-dev\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.936652 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.937062 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-nvme\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.938359 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.938449 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.935230 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.934804 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.926233 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-sys\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.937131 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-nvme\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.938746 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-run\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.940984 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.941642 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.941734 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.943584 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-run\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.943692 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-config-data\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.943817 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-lib-modules\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.943913 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.944017 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.944121 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.944234 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.944352 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/fd08b448-b737-4aa8-b263-b48e3c32f03b-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.944454 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-dns-svc\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.944579 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.944697 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.944809 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.944926 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppjtf\" (UniqueName: \"kubernetes.io/projected/4047d3e2-47ab-412e-a491-34835eac6716-kube-api-access-ppjtf\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.957776 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.957899 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.958014 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj4sb\" (UniqueName: \"kubernetes.io/projected/65ff9f68-860d-41e6-a48b-4ada29ea6275-kube-api-access-lj4sb\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.958116 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.958187 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-scripts\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.958250 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.958338 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-sys\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.949652 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-lib-modules\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.950349 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.951472 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/fd08b448-b737-4aa8-b263-b48e3c32f03b-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.953234 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.954447 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-config-data-custom\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.957725 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.959020 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/65ff9f68-860d-41e6-a48b-4ada29ea6275-ceph\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.959118 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.947116 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.959066 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.941767 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.941777 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.947145 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-run\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.948882 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.948805 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-dns-svc\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.948902 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.960076 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.960147 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-sys\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:28 crc kubenswrapper[4744]: I1003 16:44:28.960449 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.009653 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-scripts\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.010326 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.010369 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-config-data\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.010482 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.017201 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.020166 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.020813 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj4sb\" (UniqueName: \"kubernetes.io/projected/65ff9f68-860d-41e6-a48b-4ada29ea6275-kube-api-access-lj4sb\") pod \"cinder-backup-0\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " pod="openstack/cinder-backup-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.059826 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppjtf\" (UniqueName: \"kubernetes.io/projected/4047d3e2-47ab-412e-a491-34835eac6716-kube-api-access-ppjtf\") pod \"dnsmasq-dns-5784cf869f-qcrs4\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.061635 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.087622 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.090312 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.097102 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.098867 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mpmv\" (UniqueName: \"kubernetes.io/projected/fd08b448-b737-4aa8-b263-b48e3c32f03b-kube-api-access-2mpmv\") pod \"cinder-volume-volume1-0\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.104027 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.270044 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.270110 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.270157 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-config-data\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.270417 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-logs\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.270455 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-config-data-custom\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.270513 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrpg7\" (UniqueName: \"kubernetes.io/projected/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-kube-api-access-mrpg7\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.270536 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-scripts\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.353115 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.372126 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-logs\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.372177 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-config-data-custom\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.372210 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrpg7\" (UniqueName: \"kubernetes.io/projected/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-kube-api-access-mrpg7\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.372227 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-scripts\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.372258 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.372275 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.372305 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-config-data\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.373902 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-logs\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.373974 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.383264 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.387748 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.387757 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-config-data-custom\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.388650 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-scripts\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.392689 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrpg7\" (UniqueName: \"kubernetes.io/projected/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-kube-api-access-mrpg7\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.395512 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-config-data\") pod \"cinder-api-0\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.434945 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.609194 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-689c67db59-x766r"] Oct 03 16:44:29 crc kubenswrapper[4744]: I1003 16:44:29.793958 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.328776 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qcrs4"] Oct 03 16:44:30 crc kubenswrapper[4744]: W1003 16:44:30.354534 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4047d3e2_47ab_412e_a491_34835eac6716.slice/crio-18765538cc9146ab7f8bf07ac989eb94b51b91dc70cfc4c2c4d8a6bfb4a8e236 WatchSource:0}: Error finding container 18765538cc9146ab7f8bf07ac989eb94b51b91dc70cfc4c2c4d8a6bfb4a8e236: Status 404 returned error can't find the container with id 18765538cc9146ab7f8bf07ac989eb94b51b91dc70cfc4c2c4d8a6bfb4a8e236 Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.366236 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.384985 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 16:44:30 crc kubenswrapper[4744]: W1003 16:44:30.395532 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fc9bc24_b8d0_44ce_8174_4f520d39f6e2.slice/crio-e1954322b5e84ad099f25bd3c7cbdf8a382f73c3f68600e55a91af6ada28d4de WatchSource:0}: Error finding container e1954322b5e84ad099f25bd3c7cbdf8a382f73c3f68600e55a91af6ada28d4de: Status 404 returned error can't find the container with id e1954322b5e84ad099f25bd3c7cbdf8a382f73c3f68600e55a91af6ada28d4de Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.510047 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.544987 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-689c67db59-x766r" event={"ID":"1c64dd03-9cb9-402f-9697-a23538c9ebc5","Type":"ContainerStarted","Data":"6fbf84a439d189648837e8260d36bf9dc7f1d2f05774a22404922407a10d4b68"} Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.545042 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-689c67db59-x766r" event={"ID":"1c64dd03-9cb9-402f-9697-a23538c9ebc5","Type":"ContainerStarted","Data":"9b359f7924c08128048491b820bfd863c6afbe1b09959e331116fed3c210eeb5"} Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.547191 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" event={"ID":"4047d3e2-47ab-412e-a491-34835eac6716","Type":"ContainerStarted","Data":"18765538cc9146ab7f8bf07ac989eb94b51b91dc70cfc4c2c4d8a6bfb4a8e236"} Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.556400 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1","Type":"ContainerStarted","Data":"75456b9371694d2a2bfac3e7ae5718c1db9217ccac7f25b410e771f245a913db"} Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.560117 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"65ff9f68-860d-41e6-a48b-4ada29ea6275","Type":"ContainerStarted","Data":"5c3820a2d701f7dc7a943209600289be7de2c96f9df9917f39bd3aa5e4ed0347"} Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.567250 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2","Type":"ContainerStarted","Data":"e1954322b5e84ad099f25bd3c7cbdf8a382f73c3f68600e55a91af6ada28d4de"} Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.573682 4744 generic.go:334] "Generic (PLEG): container finished" podID="874834f7-8c7c-422d-9ec1-822d591f6d75" containerID="6ea220f0d199dc62a2062ec5f75b92116abceb965c5ab92e91a2b31c6d396809" exitCode=137 Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.574018 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" podUID="862d1595-46ba-4e56-8b99-2ede960d343c" containerName="dnsmasq-dns" containerID="cri-o://a1984dd5069bb5139f99baf81a8d0a6eb29a1d77b9401ddf6b28286dcd983fcc" gracePeriod=10 Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.574296 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7978488848-g4fcf" event={"ID":"874834f7-8c7c-422d-9ec1-822d591f6d75","Type":"ContainerDied","Data":"6ea220f0d199dc62a2062ec5f75b92116abceb965c5ab92e91a2b31c6d396809"} Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.578444 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.724174 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/874834f7-8c7c-422d-9ec1-822d591f6d75-scripts\") pod \"874834f7-8c7c-422d-9ec1-822d591f6d75\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.724225 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drh7c\" (UniqueName: \"kubernetes.io/projected/874834f7-8c7c-422d-9ec1-822d591f6d75-kube-api-access-drh7c\") pod \"874834f7-8c7c-422d-9ec1-822d591f6d75\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.724270 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/874834f7-8c7c-422d-9ec1-822d591f6d75-config-data\") pod \"874834f7-8c7c-422d-9ec1-822d591f6d75\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.724334 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-horizon-secret-key\") pod \"874834f7-8c7c-422d-9ec1-822d591f6d75\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.724365 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-combined-ca-bundle\") pod \"874834f7-8c7c-422d-9ec1-822d591f6d75\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.724395 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-horizon-tls-certs\") pod \"874834f7-8c7c-422d-9ec1-822d591f6d75\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.724415 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/874834f7-8c7c-422d-9ec1-822d591f6d75-logs\") pod \"874834f7-8c7c-422d-9ec1-822d591f6d75\" (UID: \"874834f7-8c7c-422d-9ec1-822d591f6d75\") " Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.726273 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/874834f7-8c7c-422d-9ec1-822d591f6d75-logs" (OuterVolumeSpecName: "logs") pod "874834f7-8c7c-422d-9ec1-822d591f6d75" (UID: "874834f7-8c7c-422d-9ec1-822d591f6d75"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.727414 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/874834f7-8c7c-422d-9ec1-822d591f6d75-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.730740 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/874834f7-8c7c-422d-9ec1-822d591f6d75-kube-api-access-drh7c" (OuterVolumeSpecName: "kube-api-access-drh7c") pod "874834f7-8c7c-422d-9ec1-822d591f6d75" (UID: "874834f7-8c7c-422d-9ec1-822d591f6d75"). InnerVolumeSpecName "kube-api-access-drh7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.755951 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "874834f7-8c7c-422d-9ec1-822d591f6d75" (UID: "874834f7-8c7c-422d-9ec1-822d591f6d75"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.811330 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/874834f7-8c7c-422d-9ec1-822d591f6d75-config-data" (OuterVolumeSpecName: "config-data") pod "874834f7-8c7c-422d-9ec1-822d591f6d75" (UID: "874834f7-8c7c-422d-9ec1-822d591f6d75"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.811858 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "874834f7-8c7c-422d-9ec1-822d591f6d75" (UID: "874834f7-8c7c-422d-9ec1-822d591f6d75"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.815975 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/874834f7-8c7c-422d-9ec1-822d591f6d75-scripts" (OuterVolumeSpecName: "scripts") pod "874834f7-8c7c-422d-9ec1-822d591f6d75" (UID: "874834f7-8c7c-422d-9ec1-822d591f6d75"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.844169 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/874834f7-8c7c-422d-9ec1-822d591f6d75-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.844208 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drh7c\" (UniqueName: \"kubernetes.io/projected/874834f7-8c7c-422d-9ec1-822d591f6d75-kube-api-access-drh7c\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.844220 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/874834f7-8c7c-422d-9ec1-822d591f6d75-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.844231 4744 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:30 crc kubenswrapper[4744]: I1003 16:44:30.844251 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.000838 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "874834f7-8c7c-422d-9ec1-822d591f6d75" (UID: "874834f7-8c7c-422d-9ec1-822d591f6d75"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.053407 4744 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/874834f7-8c7c-422d-9ec1-822d591f6d75-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.376320 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.380054 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="ceilometer-central-agent" containerID="cri-o://ed780aed8b09d2751d833e6ade94da0942fa01b3405d9f734f64467bf7ac93aa" gracePeriod=30 Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.380511 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="proxy-httpd" containerID="cri-o://df955e1affcb1641a0c07eb8a16a2befa101d98292eafdb0f18b034900de85c5" gracePeriod=30 Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.380744 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="ceilometer-notification-agent" containerID="cri-o://9b3373ebc9cf64f92e3f7fcce4bbe5d1a5fb67bb5be2f8e1a20193f4c57caabd" gracePeriod=30 Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.380807 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="sg-core" containerID="cri-o://4c28569118247e74e8bffc902ce85307d90b95985f47f1c1317d76e122f0200b" gracePeriod=30 Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.495990 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.567879 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-config\") pod \"862d1595-46ba-4e56-8b99-2ede960d343c\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.567961 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-ovsdbserver-nb\") pod \"862d1595-46ba-4e56-8b99-2ede960d343c\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.568130 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-dns-swift-storage-0\") pod \"862d1595-46ba-4e56-8b99-2ede960d343c\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.568186 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n52c\" (UniqueName: \"kubernetes.io/projected/862d1595-46ba-4e56-8b99-2ede960d343c-kube-api-access-5n52c\") pod \"862d1595-46ba-4e56-8b99-2ede960d343c\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.568214 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-ovsdbserver-sb\") pod \"862d1595-46ba-4e56-8b99-2ede960d343c\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.568235 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-dns-svc\") pod \"862d1595-46ba-4e56-8b99-2ede960d343c\" (UID: \"862d1595-46ba-4e56-8b99-2ede960d343c\") " Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.595740 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/862d1595-46ba-4e56-8b99-2ede960d343c-kube-api-access-5n52c" (OuterVolumeSpecName: "kube-api-access-5n52c") pod "862d1595-46ba-4e56-8b99-2ede960d343c" (UID: "862d1595-46ba-4e56-8b99-2ede960d343c"). InnerVolumeSpecName "kube-api-access-5n52c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.639637 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7978488848-g4fcf" event={"ID":"874834f7-8c7c-422d-9ec1-822d591f6d75","Type":"ContainerDied","Data":"5ceea66d6ac67e106ac5564e62972e6abb6747edf429191f7261cb8fbec2fecb"} Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.639686 4744 scope.go:117] "RemoveContainer" containerID="3fd2733e66910ca5adf8f373718198684ccaa21a7e01b73dbd60efbc2cae7e6c" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.640037 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.640198 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7978488848-g4fcf" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.649396 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-689c67db59-x766r" event={"ID":"1c64dd03-9cb9-402f-9697-a23538c9ebc5","Type":"ContainerStarted","Data":"29172c5c9438ad323da36a16e8e617bf2ce1c9bec47525866e353504fe4411b1"} Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.650694 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.650725 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.654856 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"fd08b448-b737-4aa8-b263-b48e3c32f03b","Type":"ContainerStarted","Data":"7b5ce915256fb24bf7666c0205a50044a4691224696a890c8c3107a5eac8c614"} Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.674823 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n52c\" (UniqueName: \"kubernetes.io/projected/862d1595-46ba-4e56-8b99-2ede960d343c-kube-api-access-5n52c\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.677270 4744 generic.go:334] "Generic (PLEG): container finished" podID="4047d3e2-47ab-412e-a491-34835eac6716" containerID="fa2048dfd084a40a56d00dda9a6959069f1c98cb6079c278fac7e4c00665658c" exitCode=0 Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.677391 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" event={"ID":"4047d3e2-47ab-412e-a491-34835eac6716","Type":"ContainerDied","Data":"fa2048dfd084a40a56d00dda9a6959069f1c98cb6079c278fac7e4c00665658c"} Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.680859 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-689c67db59-x766r" podStartSLOduration=3.680836923 podStartE2EDuration="3.680836923s" podCreationTimestamp="2025-10-03 16:44:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:31.678246125 +0000 UTC m=+1197.958122031" watchObservedRunningTime="2025-10-03 16:44:31.680836923 +0000 UTC m=+1197.960712819" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.691284 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "862d1595-46ba-4e56-8b99-2ede960d343c" (UID: "862d1595-46ba-4e56-8b99-2ede960d343c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.734238 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "862d1595-46ba-4e56-8b99-2ede960d343c" (UID: "862d1595-46ba-4e56-8b99-2ede960d343c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.744477 4744 generic.go:334] "Generic (PLEG): container finished" podID="862d1595-46ba-4e56-8b99-2ede960d343c" containerID="a1984dd5069bb5139f99baf81a8d0a6eb29a1d77b9401ddf6b28286dcd983fcc" exitCode=0 Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.744600 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" event={"ID":"862d1595-46ba-4e56-8b99-2ede960d343c","Type":"ContainerDied","Data":"a1984dd5069bb5139f99baf81a8d0a6eb29a1d77b9401ddf6b28286dcd983fcc"} Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.744635 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" event={"ID":"862d1595-46ba-4e56-8b99-2ede960d343c","Type":"ContainerDied","Data":"e9ea27c666084ce7b262c680ac76fe3bd2269bd25fba2e8175d3401135530cad"} Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.744742 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-nnv9h" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.754915 4744 generic.go:334] "Generic (PLEG): container finished" podID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerID="4c28569118247e74e8bffc902ce85307d90b95985f47f1c1317d76e122f0200b" exitCode=2 Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.755170 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f26bf136-8ce5-4187-8e27-21d73d45319e","Type":"ContainerDied","Data":"4c28569118247e74e8bffc902ce85307d90b95985f47f1c1317d76e122f0200b"} Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.765939 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7978488848-g4fcf"] Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.777535 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.777578 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.787993 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7978488848-g4fcf"] Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.849147 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "862d1595-46ba-4e56-8b99-2ede960d343c" (UID: "862d1595-46ba-4e56-8b99-2ede960d343c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.865362 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-config" (OuterVolumeSpecName: "config") pod "862d1595-46ba-4e56-8b99-2ede960d343c" (UID: "862d1595-46ba-4e56-8b99-2ede960d343c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.865384 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "862d1595-46ba-4e56-8b99-2ede960d343c" (UID: "862d1595-46ba-4e56-8b99-2ede960d343c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.880835 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.880875 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:31 crc kubenswrapper[4744]: I1003 16:44:31.880884 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/862d1595-46ba-4e56-8b99-2ede960d343c-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:32 crc kubenswrapper[4744]: I1003 16:44:32.121124 4744 scope.go:117] "RemoveContainer" containerID="6ea220f0d199dc62a2062ec5f75b92116abceb965c5ab92e91a2b31c6d396809" Oct 03 16:44:32 crc kubenswrapper[4744]: I1003 16:44:32.358580 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-nnv9h"] Oct 03 16:44:32 crc kubenswrapper[4744]: I1003 16:44:32.369112 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-nnv9h"] Oct 03 16:44:32 crc kubenswrapper[4744]: I1003 16:44:32.807071 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2","Type":"ContainerStarted","Data":"eadf8163b7c803111d7fc4b1c75a4351ab0aba00059b8aa18bf8849ded4740c8"} Oct 03 16:44:32 crc kubenswrapper[4744]: I1003 16:44:32.815017 4744 generic.go:334] "Generic (PLEG): container finished" podID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerID="df955e1affcb1641a0c07eb8a16a2befa101d98292eafdb0f18b034900de85c5" exitCode=0 Oct 03 16:44:32 crc kubenswrapper[4744]: I1003 16:44:32.815053 4744 generic.go:334] "Generic (PLEG): container finished" podID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerID="9b3373ebc9cf64f92e3f7fcce4bbe5d1a5fb67bb5be2f8e1a20193f4c57caabd" exitCode=0 Oct 03 16:44:32 crc kubenswrapper[4744]: I1003 16:44:32.815063 4744 generic.go:334] "Generic (PLEG): container finished" podID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerID="ed780aed8b09d2751d833e6ade94da0942fa01b3405d9f734f64467bf7ac93aa" exitCode=0 Oct 03 16:44:32 crc kubenswrapper[4744]: I1003 16:44:32.815115 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f26bf136-8ce5-4187-8e27-21d73d45319e","Type":"ContainerDied","Data":"df955e1affcb1641a0c07eb8a16a2befa101d98292eafdb0f18b034900de85c5"} Oct 03 16:44:32 crc kubenswrapper[4744]: I1003 16:44:32.815149 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f26bf136-8ce5-4187-8e27-21d73d45319e","Type":"ContainerDied","Data":"9b3373ebc9cf64f92e3f7fcce4bbe5d1a5fb67bb5be2f8e1a20193f4c57caabd"} Oct 03 16:44:32 crc kubenswrapper[4744]: I1003 16:44:32.815163 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f26bf136-8ce5-4187-8e27-21d73d45319e","Type":"ContainerDied","Data":"ed780aed8b09d2751d833e6ade94da0942fa01b3405d9f734f64467bf7ac93aa"} Oct 03 16:44:32 crc kubenswrapper[4744]: I1003 16:44:32.865799 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:32 crc kubenswrapper[4744]: I1003 16:44:32.868729 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1","Type":"ContainerStarted","Data":"0716d450e559e3ef093de38e99e9960e41173df8ae2deba8354f830a1a1c3623"} Oct 03 16:44:32 crc kubenswrapper[4744]: I1003 16:44:32.882636 4744 scope.go:117] "RemoveContainer" containerID="a1984dd5069bb5139f99baf81a8d0a6eb29a1d77b9401ddf6b28286dcd983fcc" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.053933 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="862d1595-46ba-4e56-8b99-2ede960d343c" path="/var/lib/kubelet/pods/862d1595-46ba-4e56-8b99-2ede960d343c/volumes" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.055119 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="874834f7-8c7c-422d-9ec1-822d591f6d75" path="/var/lib/kubelet/pods/874834f7-8c7c-422d-9ec1-822d591f6d75/volumes" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.116712 4744 scope.go:117] "RemoveContainer" containerID="2a3d0286bf2ff14c339d2d4b6446b3b9d667099252294fdf22a10fd35cf2fb60" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.343869 4744 scope.go:117] "RemoveContainer" containerID="a1984dd5069bb5139f99baf81a8d0a6eb29a1d77b9401ddf6b28286dcd983fcc" Oct 03 16:44:33 crc kubenswrapper[4744]: E1003 16:44:33.344599 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1984dd5069bb5139f99baf81a8d0a6eb29a1d77b9401ddf6b28286dcd983fcc\": container with ID starting with a1984dd5069bb5139f99baf81a8d0a6eb29a1d77b9401ddf6b28286dcd983fcc not found: ID does not exist" containerID="a1984dd5069bb5139f99baf81a8d0a6eb29a1d77b9401ddf6b28286dcd983fcc" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.344635 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1984dd5069bb5139f99baf81a8d0a6eb29a1d77b9401ddf6b28286dcd983fcc"} err="failed to get container status \"a1984dd5069bb5139f99baf81a8d0a6eb29a1d77b9401ddf6b28286dcd983fcc\": rpc error: code = NotFound desc = could not find container \"a1984dd5069bb5139f99baf81a8d0a6eb29a1d77b9401ddf6b28286dcd983fcc\": container with ID starting with a1984dd5069bb5139f99baf81a8d0a6eb29a1d77b9401ddf6b28286dcd983fcc not found: ID does not exist" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.344658 4744 scope.go:117] "RemoveContainer" containerID="2a3d0286bf2ff14c339d2d4b6446b3b9d667099252294fdf22a10fd35cf2fb60" Oct 03 16:44:33 crc kubenswrapper[4744]: E1003 16:44:33.345126 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a3d0286bf2ff14c339d2d4b6446b3b9d667099252294fdf22a10fd35cf2fb60\": container with ID starting with 2a3d0286bf2ff14c339d2d4b6446b3b9d667099252294fdf22a10fd35cf2fb60 not found: ID does not exist" containerID="2a3d0286bf2ff14c339d2d4b6446b3b9d667099252294fdf22a10fd35cf2fb60" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.345185 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a3d0286bf2ff14c339d2d4b6446b3b9d667099252294fdf22a10fd35cf2fb60"} err="failed to get container status \"2a3d0286bf2ff14c339d2d4b6446b3b9d667099252294fdf22a10fd35cf2fb60\": rpc error: code = NotFound desc = could not find container \"2a3d0286bf2ff14c339d2d4b6446b3b9d667099252294fdf22a10fd35cf2fb60\": container with ID starting with 2a3d0286bf2ff14c339d2d4b6446b3b9d667099252294fdf22a10fd35cf2fb60 not found: ID does not exist" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.475882 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.560142 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f26bf136-8ce5-4187-8e27-21d73d45319e-run-httpd\") pod \"f26bf136-8ce5-4187-8e27-21d73d45319e\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.560189 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-combined-ca-bundle\") pod \"f26bf136-8ce5-4187-8e27-21d73d45319e\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.560322 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-sg-core-conf-yaml\") pod \"f26bf136-8ce5-4187-8e27-21d73d45319e\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.560343 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-config-data\") pod \"f26bf136-8ce5-4187-8e27-21d73d45319e\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.560405 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f26bf136-8ce5-4187-8e27-21d73d45319e-log-httpd\") pod \"f26bf136-8ce5-4187-8e27-21d73d45319e\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.560505 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98n8n\" (UniqueName: \"kubernetes.io/projected/f26bf136-8ce5-4187-8e27-21d73d45319e-kube-api-access-98n8n\") pod \"f26bf136-8ce5-4187-8e27-21d73d45319e\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.560546 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-scripts\") pod \"f26bf136-8ce5-4187-8e27-21d73d45319e\" (UID: \"f26bf136-8ce5-4187-8e27-21d73d45319e\") " Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.561773 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f26bf136-8ce5-4187-8e27-21d73d45319e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f26bf136-8ce5-4187-8e27-21d73d45319e" (UID: "f26bf136-8ce5-4187-8e27-21d73d45319e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.562640 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f26bf136-8ce5-4187-8e27-21d73d45319e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f26bf136-8ce5-4187-8e27-21d73d45319e" (UID: "f26bf136-8ce5-4187-8e27-21d73d45319e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.573301 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f26bf136-8ce5-4187-8e27-21d73d45319e-kube-api-access-98n8n" (OuterVolumeSpecName: "kube-api-access-98n8n") pod "f26bf136-8ce5-4187-8e27-21d73d45319e" (UID: "f26bf136-8ce5-4187-8e27-21d73d45319e"). InnerVolumeSpecName "kube-api-access-98n8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.575332 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-scripts" (OuterVolumeSpecName: "scripts") pod "f26bf136-8ce5-4187-8e27-21d73d45319e" (UID: "f26bf136-8ce5-4187-8e27-21d73d45319e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.665854 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f26bf136-8ce5-4187-8e27-21d73d45319e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.666125 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f26bf136-8ce5-4187-8e27-21d73d45319e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.666134 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98n8n\" (UniqueName: \"kubernetes.io/projected/f26bf136-8ce5-4187-8e27-21d73d45319e-kube-api-access-98n8n\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.666144 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.677945 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f26bf136-8ce5-4187-8e27-21d73d45319e" (UID: "f26bf136-8ce5-4187-8e27-21d73d45319e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.733812 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f26bf136-8ce5-4187-8e27-21d73d45319e" (UID: "f26bf136-8ce5-4187-8e27-21d73d45319e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.767828 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.767875 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.786646 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-config-data" (OuterVolumeSpecName: "config-data") pod "f26bf136-8ce5-4187-8e27-21d73d45319e" (UID: "f26bf136-8ce5-4187-8e27-21d73d45319e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.869429 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f26bf136-8ce5-4187-8e27-21d73d45319e-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.916026 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"65ff9f68-860d-41e6-a48b-4ada29ea6275","Type":"ContainerStarted","Data":"b61091089b1f64d9c36837e9b95c7f9363f084dd5fa2df3f9e7e99746f932811"} Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.972365 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f26bf136-8ce5-4187-8e27-21d73d45319e","Type":"ContainerDied","Data":"ee5b5f0bf40d5829a8a4b5495e113bb73daf2e94887e70d3cd3a84b281a30935"} Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.972431 4744 scope.go:117] "RemoveContainer" containerID="df955e1affcb1641a0c07eb8a16a2befa101d98292eafdb0f18b034900de85c5" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.972649 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.981338 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"fd08b448-b737-4aa8-b263-b48e3c32f03b","Type":"ContainerStarted","Data":"359abd73c3de7bd2a15ad779b1c60725a6a8525c9fb24fc7a4089b2c6400b9e0"} Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.984297 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" event={"ID":"4047d3e2-47ab-412e-a491-34835eac6716","Type":"ContainerStarted","Data":"6ecd7f73349ba82f8168f463061e4ccaea47660f8e0a3feb08119ad164a7be4e"} Oct 03 16:44:33 crc kubenswrapper[4744]: I1003 16:44:33.984356 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.019061 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" podStartSLOduration=6.019039078 podStartE2EDuration="6.019039078s" podCreationTimestamp="2025-10-03 16:44:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:34.00368674 +0000 UTC m=+1200.283562636" watchObservedRunningTime="2025-10-03 16:44:34.019039078 +0000 UTC m=+1200.298914974" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.155694 4744 scope.go:117] "RemoveContainer" containerID="4c28569118247e74e8bffc902ce85307d90b95985f47f1c1317d76e122f0200b" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.162688 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.188870 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.200901 4744 scope.go:117] "RemoveContainer" containerID="9b3373ebc9cf64f92e3f7fcce4bbe5d1a5fb67bb5be2f8e1a20193f4c57caabd" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211106 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:34 crc kubenswrapper[4744]: E1003 16:44:34.211536 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="sg-core" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211548 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="sg-core" Oct 03 16:44:34 crc kubenswrapper[4744]: E1003 16:44:34.211588 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="874834f7-8c7c-422d-9ec1-822d591f6d75" containerName="horizon" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211594 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="874834f7-8c7c-422d-9ec1-822d591f6d75" containerName="horizon" Oct 03 16:44:34 crc kubenswrapper[4744]: E1003 16:44:34.211603 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="ceilometer-central-agent" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211609 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="ceilometer-central-agent" Oct 03 16:44:34 crc kubenswrapper[4744]: E1003 16:44:34.211624 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862d1595-46ba-4e56-8b99-2ede960d343c" containerName="init" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211630 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="862d1595-46ba-4e56-8b99-2ede960d343c" containerName="init" Oct 03 16:44:34 crc kubenswrapper[4744]: E1003 16:44:34.211641 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="proxy-httpd" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211646 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="proxy-httpd" Oct 03 16:44:34 crc kubenswrapper[4744]: E1003 16:44:34.211654 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="874834f7-8c7c-422d-9ec1-822d591f6d75" containerName="horizon-log" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211660 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="874834f7-8c7c-422d-9ec1-822d591f6d75" containerName="horizon-log" Oct 03 16:44:34 crc kubenswrapper[4744]: E1003 16:44:34.211673 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="ceilometer-notification-agent" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211679 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="ceilometer-notification-agent" Oct 03 16:44:34 crc kubenswrapper[4744]: E1003 16:44:34.211689 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862d1595-46ba-4e56-8b99-2ede960d343c" containerName="dnsmasq-dns" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211694 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="862d1595-46ba-4e56-8b99-2ede960d343c" containerName="dnsmasq-dns" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211851 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="proxy-httpd" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211864 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="ceilometer-central-agent" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211878 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="862d1595-46ba-4e56-8b99-2ede960d343c" containerName="dnsmasq-dns" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211896 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="ceilometer-notification-agent" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211904 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="874834f7-8c7c-422d-9ec1-822d591f6d75" containerName="horizon" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211914 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" containerName="sg-core" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.211924 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="874834f7-8c7c-422d-9ec1-822d591f6d75" containerName="horizon-log" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.213552 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.217753 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.218065 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.241778 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.301732 4744 scope.go:117] "RemoveContainer" containerID="ed780aed8b09d2751d833e6ade94da0942fa01b3405d9f734f64467bf7ac93aa" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.336608 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.336767 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68631662-30f7-459c-bacc-8d326f8b8416-log-httpd\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.337011 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-scripts\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.337082 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.337163 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-config-data\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.337296 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbf5f\" (UniqueName: \"kubernetes.io/projected/68631662-30f7-459c-bacc-8d326f8b8416-kube-api-access-pbf5f\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.337424 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68631662-30f7-459c-bacc-8d326f8b8416-run-httpd\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.439771 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68631662-30f7-459c-bacc-8d326f8b8416-log-httpd\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.440244 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-scripts\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.440375 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.440568 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-config-data\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.440732 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbf5f\" (UniqueName: \"kubernetes.io/projected/68631662-30f7-459c-bacc-8d326f8b8416-kube-api-access-pbf5f\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.440947 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68631662-30f7-459c-bacc-8d326f8b8416-run-httpd\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.441082 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.441191 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68631662-30f7-459c-bacc-8d326f8b8416-log-httpd\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.441637 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68631662-30f7-459c-bacc-8d326f8b8416-run-httpd\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.455663 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.460089 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.460800 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-config-data\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.461174 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbf5f\" (UniqueName: \"kubernetes.io/projected/68631662-30f7-459c-bacc-8d326f8b8416-kube-api-access-pbf5f\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.463263 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-scripts\") pod \"ceilometer-0\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.562803 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.668883 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.668939 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.668987 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.669814 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7b3e699f99d905cb6203d07f6a9bf0a0b414b4e9270f595c5bdaf98f6b96e23b"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.669876 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://7b3e699f99d905cb6203d07f6a9bf0a0b414b4e9270f595c5bdaf98f6b96e23b" gracePeriod=600 Oct 03 16:44:34 crc kubenswrapper[4744]: I1003 16:44:34.947727 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f26bf136-8ce5-4187-8e27-21d73d45319e" path="/var/lib/kubelet/pods/f26bf136-8ce5-4187-8e27-21d73d45319e/volumes" Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.042602 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"fd08b448-b737-4aa8-b263-b48e3c32f03b","Type":"ContainerStarted","Data":"7fb3b4a805e6dee19865f6ab7fd633ae4e6488f093a128f82810c7818bb044ea"} Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.059775 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1","Type":"ContainerStarted","Data":"dc5bd6fe3e4d8a622919d7c79c75c558165e9a11d9cbc232d117f75604e180d3"} Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.112956 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"65ff9f68-860d-41e6-a48b-4ada29ea6275","Type":"ContainerStarted","Data":"17e2d0409805a5ae4d40c9353c78ea0bda5417cc741323264f56f0eff8e26aca"} Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.156043 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=4.728623777 podStartE2EDuration="7.156018212s" podCreationTimestamp="2025-10-03 16:44:28 +0000 UTC" firstStartedPulling="2025-10-03 16:44:30.618302858 +0000 UTC m=+1196.898178754" lastFinishedPulling="2025-10-03 16:44:33.045697283 +0000 UTC m=+1199.325573189" observedRunningTime="2025-10-03 16:44:35.126696827 +0000 UTC m=+1201.406572723" watchObservedRunningTime="2025-10-03 16:44:35.156018212 +0000 UTC m=+1201.435894108" Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.165079 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2","Type":"ContainerStarted","Data":"249ea3ab890fc17b7027b4a65f534d735e48a61486bc4d64a855431adf5a67bb"} Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.165275 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" containerName="cinder-api-log" containerID="cri-o://eadf8163b7c803111d7fc4b1c75a4351ab0aba00059b8aa18bf8849ded4740c8" gracePeriod=30 Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.165570 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.165613 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" containerName="cinder-api" containerID="cri-o://249ea3ab890fc17b7027b4a65f534d735e48a61486bc4d64a855431adf5a67bb" gracePeriod=30 Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.168126 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.392205987 podStartE2EDuration="7.168112346s" podCreationTimestamp="2025-10-03 16:44:28 +0000 UTC" firstStartedPulling="2025-10-03 16:44:29.874620329 +0000 UTC m=+1196.154496225" lastFinishedPulling="2025-10-03 16:44:30.650526688 +0000 UTC m=+1196.930402584" observedRunningTime="2025-10-03 16:44:35.156183925 +0000 UTC m=+1201.436059821" watchObservedRunningTime="2025-10-03 16:44:35.168112346 +0000 UTC m=+1201.447988242" Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.194811 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="7b3e699f99d905cb6203d07f6a9bf0a0b414b4e9270f595c5bdaf98f6b96e23b" exitCode=0 Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.195050 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"7b3e699f99d905cb6203d07f6a9bf0a0b414b4e9270f595c5bdaf98f6b96e23b"} Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.195446 4744 scope.go:117] "RemoveContainer" containerID="7ad0acf841a19f4f4edda47a15b6310ad0b8ad88f8fb721a55d7cdc8e8814147" Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.196363 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=5.947307402 podStartE2EDuration="7.196351326s" podCreationTimestamp="2025-10-03 16:44:28 +0000 UTC" firstStartedPulling="2025-10-03 16:44:30.353210788 +0000 UTC m=+1196.633086684" lastFinishedPulling="2025-10-03 16:44:31.602254712 +0000 UTC m=+1197.882130608" observedRunningTime="2025-10-03 16:44:35.189419589 +0000 UTC m=+1201.469295485" watchObservedRunningTime="2025-10-03 16:44:35.196351326 +0000 UTC m=+1201.476227222" Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.225054 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.225035466 podStartE2EDuration="6.225035466s" podCreationTimestamp="2025-10-03 16:44:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:35.212887481 +0000 UTC m=+1201.492763387" watchObservedRunningTime="2025-10-03 16:44:35.225035466 +0000 UTC m=+1201.504911362" Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.313242 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.332762 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.808776 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6bc59588f-s74gr" Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.899278 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6fd7fcfb7b-kbrm2"] Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.899510 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6fd7fcfb7b-kbrm2" podUID="21f7869b-e2f2-4e15-861c-9c3df31aa03e" containerName="neutron-api" containerID="cri-o://d2cde1ba2ae3a2a6c2ad6e9d4f4fdef85ef6c41bd2856cd5e4623d1dff03487a" gracePeriod=30 Oct 03 16:44:35 crc kubenswrapper[4744]: I1003 16:44:35.899925 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6fd7fcfb7b-kbrm2" podUID="21f7869b-e2f2-4e15-861c-9c3df31aa03e" containerName="neutron-httpd" containerID="cri-o://7b0dcaeee0ce21d78370b7e5c3ab8b3d6209a881fa53dbe65f50bcc2438ff8c3" gracePeriod=30 Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.253828 4744 generic.go:334] "Generic (PLEG): container finished" podID="2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" containerID="249ea3ab890fc17b7027b4a65f534d735e48a61486bc4d64a855431adf5a67bb" exitCode=0 Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.254282 4744 generic.go:334] "Generic (PLEG): container finished" podID="2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" containerID="eadf8163b7c803111d7fc4b1c75a4351ab0aba00059b8aa18bf8849ded4740c8" exitCode=143 Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.254353 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2","Type":"ContainerDied","Data":"249ea3ab890fc17b7027b4a65f534d735e48a61486bc4d64a855431adf5a67bb"} Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.254380 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2","Type":"ContainerDied","Data":"eadf8163b7c803111d7fc4b1c75a4351ab0aba00059b8aa18bf8849ded4740c8"} Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.288789 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"60d0a3291947a22b9d553079ceec5bbbae302d529f9f4071b49d5de61562db0c"} Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.319567 4744 generic.go:334] "Generic (PLEG): container finished" podID="21f7869b-e2f2-4e15-861c-9c3df31aa03e" containerID="7b0dcaeee0ce21d78370b7e5c3ab8b3d6209a881fa53dbe65f50bcc2438ff8c3" exitCode=0 Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.320016 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fd7fcfb7b-kbrm2" event={"ID":"21f7869b-e2f2-4e15-861c-9c3df31aa03e","Type":"ContainerDied","Data":"7b0dcaeee0ce21d78370b7e5c3ab8b3d6209a881fa53dbe65f50bcc2438ff8c3"} Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.335043 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68631662-30f7-459c-bacc-8d326f8b8416","Type":"ContainerStarted","Data":"5a6ee26bfe526aa0d8dedeffe84b22d74f74207896a02f59d5ec44f4a94e362b"} Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.441560 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.605202 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-config-data\") pod \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.605485 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-scripts\") pod \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.605560 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-combined-ca-bundle\") pod \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.605607 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-etc-machine-id\") pod \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.605653 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-logs\") pod \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.605717 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-config-data-custom\") pod \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.605828 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrpg7\" (UniqueName: \"kubernetes.io/projected/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-kube-api-access-mrpg7\") pod \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\" (UID: \"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2\") " Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.606191 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" (UID: "2fc9bc24-b8d0-44ce-8174-4f520d39f6e2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.606396 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-logs" (OuterVolumeSpecName: "logs") pod "2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" (UID: "2fc9bc24-b8d0-44ce-8174-4f520d39f6e2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.613861 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" (UID: "2fc9bc24-b8d0-44ce-8174-4f520d39f6e2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.638174 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-kube-api-access-mrpg7" (OuterVolumeSpecName: "kube-api-access-mrpg7") pod "2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" (UID: "2fc9bc24-b8d0-44ce-8174-4f520d39f6e2"). InnerVolumeSpecName "kube-api-access-mrpg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.663011 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" (UID: "2fc9bc24-b8d0-44ce-8174-4f520d39f6e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.663097 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-scripts" (OuterVolumeSpecName: "scripts") pod "2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" (UID: "2fc9bc24-b8d0-44ce-8174-4f520d39f6e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.708610 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.708646 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.708655 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.708664 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrpg7\" (UniqueName: \"kubernetes.io/projected/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-kube-api-access-mrpg7\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.708675 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.708683 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.715936 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-config-data" (OuterVolumeSpecName: "config-data") pod "2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" (UID: "2fc9bc24-b8d0-44ce-8174-4f520d39f6e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:36 crc kubenswrapper[4744]: I1003 16:44:36.811125 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.182423 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.377372 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2fc9bc24-b8d0-44ce-8174-4f520d39f6e2","Type":"ContainerDied","Data":"e1954322b5e84ad099f25bd3c7cbdf8a382f73c3f68600e55a91af6ada28d4de"} Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.377594 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.377714 4744 scope.go:117] "RemoveContainer" containerID="249ea3ab890fc17b7027b4a65f534d735e48a61486bc4d64a855431adf5a67bb" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.385692 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68631662-30f7-459c-bacc-8d326f8b8416","Type":"ContainerStarted","Data":"a715f005eeddefc330647709836f3e4428fd6bf79cdba4079f4dda6559cfcf68"} Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.385741 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68631662-30f7-459c-bacc-8d326f8b8416","Type":"ContainerStarted","Data":"64e1d8fa67bd4b27a329c86f45dd74b6928ba9f3ae1ce41f3a1deadae7090fa2"} Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.409371 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.433657 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.435305 4744 scope.go:117] "RemoveContainer" containerID="eadf8163b7c803111d7fc4b1c75a4351ab0aba00059b8aa18bf8849ded4740c8" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.529186 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 03 16:44:37 crc kubenswrapper[4744]: E1003 16:44:37.530519 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" containerName="cinder-api-log" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.530684 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" containerName="cinder-api-log" Oct 03 16:44:37 crc kubenswrapper[4744]: E1003 16:44:37.530791 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" containerName="cinder-api" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.530859 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" containerName="cinder-api" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.531708 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" containerName="cinder-api-log" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.531814 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" containerName="cinder-api" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.534570 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.537792 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.539250 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.540421 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.575296 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.630815 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5j6q\" (UniqueName: \"kubernetes.io/projected/cee08598-c007-4049-8b75-d4be9c1b0796-kube-api-access-s5j6q\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.630883 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.630939 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-public-tls-certs\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.630963 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cee08598-c007-4049-8b75-d4be9c1b0796-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.630984 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-scripts\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.631025 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.631051 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cee08598-c007-4049-8b75-d4be9c1b0796-logs\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.631079 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-config-data-custom\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.631223 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-config-data\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.732859 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-public-tls-certs\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.732904 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cee08598-c007-4049-8b75-d4be9c1b0796-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.732920 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-scripts\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.732956 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.732976 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cee08598-c007-4049-8b75-d4be9c1b0796-logs\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.732995 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-config-data-custom\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.733088 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-config-data\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.733139 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5j6q\" (UniqueName: \"kubernetes.io/projected/cee08598-c007-4049-8b75-d4be9c1b0796-kube-api-access-s5j6q\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.733160 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.733831 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cee08598-c007-4049-8b75-d4be9c1b0796-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.733946 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cee08598-c007-4049-8b75-d4be9c1b0796-logs\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.738272 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.738656 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-public-tls-certs\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.741535 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-config-data\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.745024 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-scripts\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.745933 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-config-data-custom\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.750346 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee08598-c007-4049-8b75-d4be9c1b0796-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.772320 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5j6q\" (UniqueName: \"kubernetes.io/projected/cee08598-c007-4049-8b75-d4be9c1b0796-kube-api-access-s5j6q\") pod \"cinder-api-0\" (UID: \"cee08598-c007-4049-8b75-d4be9c1b0796\") " pod="openstack/cinder-api-0" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.960282 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7bf8c46bb6-7m9v6" Oct 03 16:44:37 crc kubenswrapper[4744]: I1003 16:44:37.975739 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 16:44:38 crc kubenswrapper[4744]: I1003 16:44:38.103452 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5467484d76-wdx4m"] Oct 03 16:44:38 crc kubenswrapper[4744]: I1003 16:44:38.103999 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 16:44:38 crc kubenswrapper[4744]: I1003 16:44:38.104118 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5467484d76-wdx4m" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api-log" containerID="cri-o://8a8ff0666730ce0e751b72083f365fc1952aa3b0d6aeba77566dba6aead408c9" gracePeriod=30 Oct 03 16:44:38 crc kubenswrapper[4744]: I1003 16:44:38.104320 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5467484d76-wdx4m" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api" containerID="cri-o://96bf8b18b0a9bf7861ee3471bb012ef363c641506d0431562a761c26266e3de2" gracePeriod=30 Oct 03 16:44:38 crc kubenswrapper[4744]: I1003 16:44:38.125787 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5467484d76-wdx4m" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": EOF" Oct 03 16:44:38 crc kubenswrapper[4744]: I1003 16:44:38.132307 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5467484d76-wdx4m" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": EOF" Oct 03 16:44:38 crc kubenswrapper[4744]: I1003 16:44:38.132719 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5467484d76-wdx4m" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": EOF" Oct 03 16:44:38 crc kubenswrapper[4744]: I1003 16:44:38.438154 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68631662-30f7-459c-bacc-8d326f8b8416","Type":"ContainerStarted","Data":"dc43d1e8772042d61a68de1f2b1c1c2ab881e61eac2e180894cd70a3c7153adf"} Oct 03 16:44:38 crc kubenswrapper[4744]: I1003 16:44:38.447453 4744 generic.go:334] "Generic (PLEG): container finished" podID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerID="8a8ff0666730ce0e751b72083f365fc1952aa3b0d6aeba77566dba6aead408c9" exitCode=143 Oct 03 16:44:38 crc kubenswrapper[4744]: I1003 16:44:38.447502 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5467484d76-wdx4m" event={"ID":"102bbbfd-8dd8-4946-9f35-a22a87ebaaed","Type":"ContainerDied","Data":"8a8ff0666730ce0e751b72083f365fc1952aa3b0d6aeba77566dba6aead408c9"} Oct 03 16:44:38 crc kubenswrapper[4744]: I1003 16:44:38.582095 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:38 crc kubenswrapper[4744]: I1003 16:44:38.584319 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-689c67db59-x766r" Oct 03 16:44:38 crc kubenswrapper[4744]: I1003 16:44:38.744460 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 16:44:38 crc kubenswrapper[4744]: W1003 16:44:38.745613 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcee08598_c007_4049_8b75_d4be9c1b0796.slice/crio-204c94d7d1c827cc69b5ee61e696c30122d1d566ad8b7c86903e7dbd6f78dbd2 WatchSource:0}: Error finding container 204c94d7d1c827cc69b5ee61e696c30122d1d566ad8b7c86903e7dbd6f78dbd2: Status 404 returned error can't find the container with id 204c94d7d1c827cc69b5ee61e696c30122d1d566ad8b7c86903e7dbd6f78dbd2 Oct 03 16:44:38 crc kubenswrapper[4744]: I1003 16:44:38.919049 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fc9bc24-b8d0-44ce-8174-4f520d39f6e2" path="/var/lib/kubelet/pods/2fc9bc24-b8d0-44ce-8174-4f520d39f6e2/volumes" Oct 03 16:44:39 crc kubenswrapper[4744]: I1003 16:44:39.011380 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 03 16:44:39 crc kubenswrapper[4744]: I1003 16:44:39.063418 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 03 16:44:39 crc kubenswrapper[4744]: I1003 16:44:39.297120 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-backup-0" podUID="65ff9f68-860d-41e6-a48b-4ada29ea6275" containerName="cinder-backup" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 16:44:39 crc kubenswrapper[4744]: I1003 16:44:39.355114 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:39 crc kubenswrapper[4744]: I1003 16:44:39.385849 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:39 crc kubenswrapper[4744]: I1003 16:44:39.441777 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-6bfh4"] Oct 03 16:44:39 crc kubenswrapper[4744]: I1003 16:44:39.442484 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" podUID="4482beb6-50d9-4c37-8b51-927dbacbc455" containerName="dnsmasq-dns" containerID="cri-o://ce53319130146b9d4474d482dcd8399ab7aac5d448defefe0fe4d9f643c1484d" gracePeriod=10 Oct 03 16:44:39 crc kubenswrapper[4744]: I1003 16:44:39.468628 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cee08598-c007-4049-8b75-d4be9c1b0796","Type":"ContainerStarted","Data":"204c94d7d1c827cc69b5ee61e696c30122d1d566ad8b7c86903e7dbd6f78dbd2"} Oct 03 16:44:39 crc kubenswrapper[4744]: I1003 16:44:39.608854 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 03 16:44:39 crc kubenswrapper[4744]: I1003 16:44:39.693051 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 16:44:39 crc kubenswrapper[4744]: I1003 16:44:39.806879 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:39 crc kubenswrapper[4744]: I1003 16:44:39.849599 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.267279 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.406222 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-ovsdbserver-nb\") pod \"4482beb6-50d9-4c37-8b51-927dbacbc455\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.406293 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-dns-svc\") pod \"4482beb6-50d9-4c37-8b51-927dbacbc455\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.406325 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-config\") pod \"4482beb6-50d9-4c37-8b51-927dbacbc455\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.406495 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcjtd\" (UniqueName: \"kubernetes.io/projected/4482beb6-50d9-4c37-8b51-927dbacbc455-kube-api-access-rcjtd\") pod \"4482beb6-50d9-4c37-8b51-927dbacbc455\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.406567 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-ovsdbserver-sb\") pod \"4482beb6-50d9-4c37-8b51-927dbacbc455\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.406619 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-dns-swift-storage-0\") pod \"4482beb6-50d9-4c37-8b51-927dbacbc455\" (UID: \"4482beb6-50d9-4c37-8b51-927dbacbc455\") " Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.437372 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4482beb6-50d9-4c37-8b51-927dbacbc455-kube-api-access-rcjtd" (OuterVolumeSpecName: "kube-api-access-rcjtd") pod "4482beb6-50d9-4c37-8b51-927dbacbc455" (UID: "4482beb6-50d9-4c37-8b51-927dbacbc455"). InnerVolumeSpecName "kube-api-access-rcjtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.476096 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4482beb6-50d9-4c37-8b51-927dbacbc455" (UID: "4482beb6-50d9-4c37-8b51-927dbacbc455"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.481762 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4482beb6-50d9-4c37-8b51-927dbacbc455" (UID: "4482beb6-50d9-4c37-8b51-927dbacbc455"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.505943 4744 generic.go:334] "Generic (PLEG): container finished" podID="df7dff0d-99e0-4e98-b752-9096d24e314b" containerID="e7656f6bbb2b5a8116c6daa66427fab681e8be77ee5904e4f595b06426e76e0e" exitCode=0 Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.506048 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-8bj8c" event={"ID":"df7dff0d-99e0-4e98-b752-9096d24e314b","Type":"ContainerDied","Data":"e7656f6bbb2b5a8116c6daa66427fab681e8be77ee5904e4f595b06426e76e0e"} Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.507715 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4482beb6-50d9-4c37-8b51-927dbacbc455" (UID: "4482beb6-50d9-4c37-8b51-927dbacbc455"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.508984 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.508998 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.509009 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.509019 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcjtd\" (UniqueName: \"kubernetes.io/projected/4482beb6-50d9-4c37-8b51-927dbacbc455-kube-api-access-rcjtd\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.525387 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-config" (OuterVolumeSpecName: "config") pod "4482beb6-50d9-4c37-8b51-927dbacbc455" (UID: "4482beb6-50d9-4c37-8b51-927dbacbc455"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.553945 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4482beb6-50d9-4c37-8b51-927dbacbc455" (UID: "4482beb6-50d9-4c37-8b51-927dbacbc455"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.567406 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68631662-30f7-459c-bacc-8d326f8b8416","Type":"ContainerStarted","Data":"ba6406170f7106c9dba94384d81f63d45eb970af3d62d53c86e31b1dc85d47ff"} Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.567732 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.579004 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cee08598-c007-4049-8b75-d4be9c1b0796","Type":"ContainerStarted","Data":"cd137dd07ba13e005b51b6c6a9dac9dbf1444564a994eacad765328a3ba231e0"} Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.585202 4744 generic.go:334] "Generic (PLEG): container finished" podID="4482beb6-50d9-4c37-8b51-927dbacbc455" containerID="ce53319130146b9d4474d482dcd8399ab7aac5d448defefe0fe4d9f643c1484d" exitCode=0 Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.585275 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" event={"ID":"4482beb6-50d9-4c37-8b51-927dbacbc455","Type":"ContainerDied","Data":"ce53319130146b9d4474d482dcd8399ab7aac5d448defefe0fe4d9f643c1484d"} Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.585321 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" event={"ID":"4482beb6-50d9-4c37-8b51-927dbacbc455","Type":"ContainerDied","Data":"074afbc4c4fff049b90bb08afdf33f4a9d95b28878621ecb42443352e4586732"} Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.585407 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" containerName="cinder-scheduler" containerID="cri-o://0716d450e559e3ef093de38e99e9960e41173df8ae2deba8354f830a1a1c3623" gracePeriod=30 Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.585595 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-volume-volume1-0" podUID="fd08b448-b737-4aa8-b263-b48e3c32f03b" containerName="cinder-volume" containerID="cri-o://359abd73c3de7bd2a15ad779b1c60725a6a8525c9fb24fc7a4089b2c6400b9e0" gracePeriod=30 Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.585880 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" containerName="probe" containerID="cri-o://dc5bd6fe3e4d8a622919d7c79c75c558165e9a11d9cbc232d117f75604e180d3" gracePeriod=30 Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.586076 4744 scope.go:117] "RemoveContainer" containerID="ce53319130146b9d4474d482dcd8399ab7aac5d448defefe0fe4d9f643c1484d" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.586310 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-6bfh4" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.586062 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-volume-volume1-0" podUID="fd08b448-b737-4aa8-b263-b48e3c32f03b" containerName="probe" containerID="cri-o://7fb3b4a805e6dee19865f6ab7fd633ae4e6488f093a128f82810c7818bb044ea" gracePeriod=30 Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.598705 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.694306733 podStartE2EDuration="6.598688611s" podCreationTimestamp="2025-10-03 16:44:34 +0000 UTC" firstStartedPulling="2025-10-03 16:44:35.339320537 +0000 UTC m=+1201.619196433" lastFinishedPulling="2025-10-03 16:44:39.243702405 +0000 UTC m=+1205.523578311" observedRunningTime="2025-10-03 16:44:40.584573421 +0000 UTC m=+1206.864449317" watchObservedRunningTime="2025-10-03 16:44:40.598688611 +0000 UTC m=+1206.878564507" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.610645 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.610667 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4482beb6-50d9-4c37-8b51-927dbacbc455-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.635990 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-6bfh4"] Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.644985 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-6bfh4"] Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.646366 4744 scope.go:117] "RemoveContainer" containerID="3b70138207949fd8d39b629b71a2550352f9a5faa2d9ce32c6a3c59417f6993b" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.728161 4744 scope.go:117] "RemoveContainer" containerID="ce53319130146b9d4474d482dcd8399ab7aac5d448defefe0fe4d9f643c1484d" Oct 03 16:44:40 crc kubenswrapper[4744]: E1003 16:44:40.728700 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce53319130146b9d4474d482dcd8399ab7aac5d448defefe0fe4d9f643c1484d\": container with ID starting with ce53319130146b9d4474d482dcd8399ab7aac5d448defefe0fe4d9f643c1484d not found: ID does not exist" containerID="ce53319130146b9d4474d482dcd8399ab7aac5d448defefe0fe4d9f643c1484d" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.728735 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce53319130146b9d4474d482dcd8399ab7aac5d448defefe0fe4d9f643c1484d"} err="failed to get container status \"ce53319130146b9d4474d482dcd8399ab7aac5d448defefe0fe4d9f643c1484d\": rpc error: code = NotFound desc = could not find container \"ce53319130146b9d4474d482dcd8399ab7aac5d448defefe0fe4d9f643c1484d\": container with ID starting with ce53319130146b9d4474d482dcd8399ab7aac5d448defefe0fe4d9f643c1484d not found: ID does not exist" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.728756 4744 scope.go:117] "RemoveContainer" containerID="3b70138207949fd8d39b629b71a2550352f9a5faa2d9ce32c6a3c59417f6993b" Oct 03 16:44:40 crc kubenswrapper[4744]: E1003 16:44:40.729449 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b70138207949fd8d39b629b71a2550352f9a5faa2d9ce32c6a3c59417f6993b\": container with ID starting with 3b70138207949fd8d39b629b71a2550352f9a5faa2d9ce32c6a3c59417f6993b not found: ID does not exist" containerID="3b70138207949fd8d39b629b71a2550352f9a5faa2d9ce32c6a3c59417f6993b" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.729496 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b70138207949fd8d39b629b71a2550352f9a5faa2d9ce32c6a3c59417f6993b"} err="failed to get container status \"3b70138207949fd8d39b629b71a2550352f9a5faa2d9ce32c6a3c59417f6993b\": rpc error: code = NotFound desc = could not find container \"3b70138207949fd8d39b629b71a2550352f9a5faa2d9ce32c6a3c59417f6993b\": container with ID starting with 3b70138207949fd8d39b629b71a2550352f9a5faa2d9ce32c6a3c59417f6993b not found: ID does not exist" Oct 03 16:44:40 crc kubenswrapper[4744]: I1003 16:44:40.920065 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4482beb6-50d9-4c37-8b51-927dbacbc455" path="/var/lib/kubelet/pods/4482beb6-50d9-4c37-8b51-927dbacbc455/volumes" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.411221 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.538805 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-combined-ca-bundle\") pod \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.539129 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-ovndb-tls-certs\") pod \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.539315 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqnrn\" (UniqueName: \"kubernetes.io/projected/21f7869b-e2f2-4e15-861c-9c3df31aa03e-kube-api-access-xqnrn\") pod \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.539359 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-httpd-config\") pod \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.539375 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-config\") pod \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\" (UID: \"21f7869b-e2f2-4e15-861c-9c3df31aa03e\") " Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.545044 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21f7869b-e2f2-4e15-861c-9c3df31aa03e-kube-api-access-xqnrn" (OuterVolumeSpecName: "kube-api-access-xqnrn") pod "21f7869b-e2f2-4e15-861c-9c3df31aa03e" (UID: "21f7869b-e2f2-4e15-861c-9c3df31aa03e"). InnerVolumeSpecName "kube-api-access-xqnrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.545794 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "21f7869b-e2f2-4e15-861c-9c3df31aa03e" (UID: "21f7869b-e2f2-4e15-861c-9c3df31aa03e"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.631786 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21f7869b-e2f2-4e15-861c-9c3df31aa03e" (UID: "21f7869b-e2f2-4e15-861c-9c3df31aa03e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.641754 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.641800 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.641816 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqnrn\" (UniqueName: \"kubernetes.io/projected/21f7869b-e2f2-4e15-861c-9c3df31aa03e-kube-api-access-xqnrn\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.646599 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-config" (OuterVolumeSpecName: "config") pod "21f7869b-e2f2-4e15-861c-9c3df31aa03e" (UID: "21f7869b-e2f2-4e15-861c-9c3df31aa03e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.647189 4744 generic.go:334] "Generic (PLEG): container finished" podID="21f7869b-e2f2-4e15-861c-9c3df31aa03e" containerID="d2cde1ba2ae3a2a6c2ad6e9d4f4fdef85ef6c41bd2856cd5e4623d1dff03487a" exitCode=0 Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.647256 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fd7fcfb7b-kbrm2" event={"ID":"21f7869b-e2f2-4e15-861c-9c3df31aa03e","Type":"ContainerDied","Data":"d2cde1ba2ae3a2a6c2ad6e9d4f4fdef85ef6c41bd2856cd5e4623d1dff03487a"} Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.647290 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6fd7fcfb7b-kbrm2" event={"ID":"21f7869b-e2f2-4e15-861c-9c3df31aa03e","Type":"ContainerDied","Data":"bbc5a489ad2c5f3da6cc3beb9575ad8ad8298e98dda6a843c47746732300e126"} Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.647311 4744 scope.go:117] "RemoveContainer" containerID="7b0dcaeee0ce21d78370b7e5c3ab8b3d6209a881fa53dbe65f50bcc2438ff8c3" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.647404 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6fd7fcfb7b-kbrm2" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.657004 4744 generic.go:334] "Generic (PLEG): container finished" podID="fd08b448-b737-4aa8-b263-b48e3c32f03b" containerID="7fb3b4a805e6dee19865f6ab7fd633ae4e6488f093a128f82810c7818bb044ea" exitCode=0 Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.659773 4744 generic.go:334] "Generic (PLEG): container finished" podID="fd08b448-b737-4aa8-b263-b48e3c32f03b" containerID="359abd73c3de7bd2a15ad779b1c60725a6a8525c9fb24fc7a4089b2c6400b9e0" exitCode=0 Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.656981 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"fd08b448-b737-4aa8-b263-b48e3c32f03b","Type":"ContainerDied","Data":"7fb3b4a805e6dee19865f6ab7fd633ae4e6488f093a128f82810c7818bb044ea"} Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.659858 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"fd08b448-b737-4aa8-b263-b48e3c32f03b","Type":"ContainerDied","Data":"359abd73c3de7bd2a15ad779b1c60725a6a8525c9fb24fc7a4089b2c6400b9e0"} Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.659933 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "21f7869b-e2f2-4e15-861c-9c3df31aa03e" (UID: "21f7869b-e2f2-4e15-861c-9c3df31aa03e"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.665271 4744 generic.go:334] "Generic (PLEG): container finished" podID="4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" containerID="0716d450e559e3ef093de38e99e9960e41173df8ae2deba8354f830a1a1c3623" exitCode=0 Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.665321 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1","Type":"ContainerDied","Data":"0716d450e559e3ef093de38e99e9960e41173df8ae2deba8354f830a1a1c3623"} Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.670329 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cee08598-c007-4049-8b75-d4be9c1b0796","Type":"ContainerStarted","Data":"f393fe5fa8fb48094732dd41da53b2676b270684f9a79b6a56d1263382055e5a"} Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.670540 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.671387 4744 scope.go:117] "RemoveContainer" containerID="d2cde1ba2ae3a2a6c2ad6e9d4f4fdef85ef6c41bd2856cd5e4623d1dff03487a" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.698972 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.6989403119999995 podStartE2EDuration="4.698940312s" podCreationTimestamp="2025-10-03 16:44:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:41.689033137 +0000 UTC m=+1207.968909033" watchObservedRunningTime="2025-10-03 16:44:41.698940312 +0000 UTC m=+1207.978816208" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.715093 4744 scope.go:117] "RemoveContainer" containerID="7b0dcaeee0ce21d78370b7e5c3ab8b3d6209a881fa53dbe65f50bcc2438ff8c3" Oct 03 16:44:41 crc kubenswrapper[4744]: E1003 16:44:41.718588 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b0dcaeee0ce21d78370b7e5c3ab8b3d6209a881fa53dbe65f50bcc2438ff8c3\": container with ID starting with 7b0dcaeee0ce21d78370b7e5c3ab8b3d6209a881fa53dbe65f50bcc2438ff8c3 not found: ID does not exist" containerID="7b0dcaeee0ce21d78370b7e5c3ab8b3d6209a881fa53dbe65f50bcc2438ff8c3" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.718615 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b0dcaeee0ce21d78370b7e5c3ab8b3d6209a881fa53dbe65f50bcc2438ff8c3"} err="failed to get container status \"7b0dcaeee0ce21d78370b7e5c3ab8b3d6209a881fa53dbe65f50bcc2438ff8c3\": rpc error: code = NotFound desc = could not find container \"7b0dcaeee0ce21d78370b7e5c3ab8b3d6209a881fa53dbe65f50bcc2438ff8c3\": container with ID starting with 7b0dcaeee0ce21d78370b7e5c3ab8b3d6209a881fa53dbe65f50bcc2438ff8c3 not found: ID does not exist" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.718635 4744 scope.go:117] "RemoveContainer" containerID="d2cde1ba2ae3a2a6c2ad6e9d4f4fdef85ef6c41bd2856cd5e4623d1dff03487a" Oct 03 16:44:41 crc kubenswrapper[4744]: E1003 16:44:41.722568 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2cde1ba2ae3a2a6c2ad6e9d4f4fdef85ef6c41bd2856cd5e4623d1dff03487a\": container with ID starting with d2cde1ba2ae3a2a6c2ad6e9d4f4fdef85ef6c41bd2856cd5e4623d1dff03487a not found: ID does not exist" containerID="d2cde1ba2ae3a2a6c2ad6e9d4f4fdef85ef6c41bd2856cd5e4623d1dff03487a" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.722592 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2cde1ba2ae3a2a6c2ad6e9d4f4fdef85ef6c41bd2856cd5e4623d1dff03487a"} err="failed to get container status \"d2cde1ba2ae3a2a6c2ad6e9d4f4fdef85ef6c41bd2856cd5e4623d1dff03487a\": rpc error: code = NotFound desc = could not find container \"d2cde1ba2ae3a2a6c2ad6e9d4f4fdef85ef6c41bd2856cd5e4623d1dff03487a\": container with ID starting with d2cde1ba2ae3a2a6c2ad6e9d4f4fdef85ef6c41bd2856cd5e4623d1dff03487a not found: ID does not exist" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.747249 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.747297 4744 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/21f7869b-e2f2-4e15-861c-9c3df31aa03e-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:41 crc kubenswrapper[4744]: I1003 16:44:41.978767 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.117696 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6fd7fcfb7b-kbrm2"] Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.126739 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6fd7fcfb7b-kbrm2"] Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164153 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-run\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164205 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-scripts\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164242 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-locks-cinder\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164288 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-config-data-custom\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164325 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-machine-id\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164347 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-config-data\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164391 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-lib-cinder\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164406 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-nvme\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164467 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mpmv\" (UniqueName: \"kubernetes.io/projected/fd08b448-b737-4aa8-b263-b48e3c32f03b-kube-api-access-2mpmv\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164497 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-locks-brick\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164547 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-dev\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164569 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/fd08b448-b737-4aa8-b263-b48e3c32f03b-ceph\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164594 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-sys\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164614 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-lib-modules\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164656 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-combined-ca-bundle\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164678 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-iscsi\") pod \"fd08b448-b737-4aa8-b263-b48e3c32f03b\" (UID: \"fd08b448-b737-4aa8-b263-b48e3c32f03b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.164997 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.165042 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.165103 4744 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.165128 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.169043 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.169117 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.169138 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-sys" (OuterVolumeSpecName: "sys") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.170972 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd08b448-b737-4aa8-b263-b48e3c32f03b-kube-api-access-2mpmv" (OuterVolumeSpecName: "kube-api-access-2mpmv") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "kube-api-access-2mpmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.171051 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.171082 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-run" (OuterVolumeSpecName: "run") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.171108 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.171139 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-dev" (OuterVolumeSpecName: "dev") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.172823 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.177994 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-scripts" (OuterVolumeSpecName: "scripts") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.179541 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd08b448-b737-4aa8-b263-b48e3c32f03b-ceph" (OuterVolumeSpecName: "ceph") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.214758 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.222147 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-8bj8c" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.266337 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.268019 4744 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.268052 4744 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-dev\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.268065 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/fd08b448-b737-4aa8-b263-b48e3c32f03b-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.268079 4744 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-sys\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.268115 4744 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.268126 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.268137 4744 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.268147 4744 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-run\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.268158 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.268168 4744 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.268178 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.268188 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.268200 4744 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/fd08b448-b737-4aa8-b263-b48e3c32f03b-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.268209 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mpmv\" (UniqueName: \"kubernetes.io/projected/fd08b448-b737-4aa8-b263-b48e3c32f03b-kube-api-access-2mpmv\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.328018 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-config-data" (OuterVolumeSpecName: "config-data") pod "fd08b448-b737-4aa8-b263-b48e3c32f03b" (UID: "fd08b448-b737-4aa8-b263-b48e3c32f03b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.369629 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58hb9\" (UniqueName: \"kubernetes.io/projected/df7dff0d-99e0-4e98-b752-9096d24e314b-kube-api-access-58hb9\") pod \"df7dff0d-99e0-4e98-b752-9096d24e314b\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.369737 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-scripts\") pod \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.369795 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-job-config-data\") pod \"df7dff0d-99e0-4e98-b752-9096d24e314b\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.369883 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-combined-ca-bundle\") pod \"df7dff0d-99e0-4e98-b752-9096d24e314b\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.369940 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-config-data-custom\") pod \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.370718 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-combined-ca-bundle\") pod \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.370753 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-config-data\") pod \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.370785 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-etc-machine-id\") pod \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.370803 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgpfl\" (UniqueName: \"kubernetes.io/projected/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-kube-api-access-mgpfl\") pod \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\" (UID: \"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.370854 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-config-data\") pod \"df7dff0d-99e0-4e98-b752-9096d24e314b\" (UID: \"df7dff0d-99e0-4e98-b752-9096d24e314b\") " Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.370918 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" (UID: "4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.371717 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd08b448-b737-4aa8-b263-b48e3c32f03b-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.371746 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.373840 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "df7dff0d-99e0-4e98-b752-9096d24e314b" (UID: "df7dff0d-99e0-4e98-b752-9096d24e314b"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.374114 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-kube-api-access-mgpfl" (OuterVolumeSpecName: "kube-api-access-mgpfl") pod "4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" (UID: "4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1"). InnerVolumeSpecName "kube-api-access-mgpfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.378159 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-scripts" (OuterVolumeSpecName: "scripts") pod "4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" (UID: "4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.380429 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" (UID: "4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.383443 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df7dff0d-99e0-4e98-b752-9096d24e314b-kube-api-access-58hb9" (OuterVolumeSpecName: "kube-api-access-58hb9") pod "df7dff0d-99e0-4e98-b752-9096d24e314b" (UID: "df7dff0d-99e0-4e98-b752-9096d24e314b"). InnerVolumeSpecName "kube-api-access-58hb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.386243 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-config-data" (OuterVolumeSpecName: "config-data") pod "df7dff0d-99e0-4e98-b752-9096d24e314b" (UID: "df7dff0d-99e0-4e98-b752-9096d24e314b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.410740 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df7dff0d-99e0-4e98-b752-9096d24e314b" (UID: "df7dff0d-99e0-4e98-b752-9096d24e314b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.427941 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" (UID: "4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.473598 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgpfl\" (UniqueName: \"kubernetes.io/projected/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-kube-api-access-mgpfl\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.473642 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.473657 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58hb9\" (UniqueName: \"kubernetes.io/projected/df7dff0d-99e0-4e98-b752-9096d24e314b-kube-api-access-58hb9\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.473670 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.473682 4744 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.473693 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df7dff0d-99e0-4e98-b752-9096d24e314b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.473704 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.473715 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.500413 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-config-data" (OuterVolumeSpecName: "config-data") pod "4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" (UID: "4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.575554 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.692436 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 03 16:44:42 crc kubenswrapper[4744]: E1003 16:44:42.692821 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd08b448-b737-4aa8-b263-b48e3c32f03b" containerName="probe" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.692837 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd08b448-b737-4aa8-b263-b48e3c32f03b" containerName="probe" Oct 03 16:44:42 crc kubenswrapper[4744]: E1003 16:44:42.692848 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd08b448-b737-4aa8-b263-b48e3c32f03b" containerName="cinder-volume" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.692853 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd08b448-b737-4aa8-b263-b48e3c32f03b" containerName="cinder-volume" Oct 03 16:44:42 crc kubenswrapper[4744]: E1003 16:44:42.692866 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df7dff0d-99e0-4e98-b752-9096d24e314b" containerName="manila-db-sync" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.692872 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="df7dff0d-99e0-4e98-b752-9096d24e314b" containerName="manila-db-sync" Oct 03 16:44:42 crc kubenswrapper[4744]: E1003 16:44:42.692885 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4482beb6-50d9-4c37-8b51-927dbacbc455" containerName="dnsmasq-dns" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.692891 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4482beb6-50d9-4c37-8b51-927dbacbc455" containerName="dnsmasq-dns" Oct 03 16:44:42 crc kubenswrapper[4744]: E1003 16:44:42.692902 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4482beb6-50d9-4c37-8b51-927dbacbc455" containerName="init" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.692908 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4482beb6-50d9-4c37-8b51-927dbacbc455" containerName="init" Oct 03 16:44:42 crc kubenswrapper[4744]: E1003 16:44:42.692920 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21f7869b-e2f2-4e15-861c-9c3df31aa03e" containerName="neutron-api" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.692926 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="21f7869b-e2f2-4e15-861c-9c3df31aa03e" containerName="neutron-api" Oct 03 16:44:42 crc kubenswrapper[4744]: E1003 16:44:42.692941 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" containerName="probe" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.692949 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" containerName="probe" Oct 03 16:44:42 crc kubenswrapper[4744]: E1003 16:44:42.692968 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" containerName="cinder-scheduler" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.692973 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" containerName="cinder-scheduler" Oct 03 16:44:42 crc kubenswrapper[4744]: E1003 16:44:42.692984 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21f7869b-e2f2-4e15-861c-9c3df31aa03e" containerName="neutron-httpd" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.692989 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="21f7869b-e2f2-4e15-861c-9c3df31aa03e" containerName="neutron-httpd" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.693156 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="21f7869b-e2f2-4e15-861c-9c3df31aa03e" containerName="neutron-api" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.693171 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd08b448-b737-4aa8-b263-b48e3c32f03b" containerName="probe" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.693181 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="df7dff0d-99e0-4e98-b752-9096d24e314b" containerName="manila-db-sync" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.693192 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4482beb6-50d9-4c37-8b51-927dbacbc455" containerName="dnsmasq-dns" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.693207 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="21f7869b-e2f2-4e15-861c-9c3df31aa03e" containerName="neutron-httpd" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.693220 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd08b448-b737-4aa8-b263-b48e3c32f03b" containerName="cinder-volume" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.693227 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" containerName="cinder-scheduler" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.693234 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" containerName="probe" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.694128 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.698057 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.698306 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.711048 4744 generic.go:334] "Generic (PLEG): container finished" podID="4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" containerID="dc5bd6fe3e4d8a622919d7c79c75c558165e9a11d9cbc232d117f75604e180d3" exitCode=0 Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.711316 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1","Type":"ContainerDied","Data":"dc5bd6fe3e4d8a622919d7c79c75c558165e9a11d9cbc232d117f75604e180d3"} Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.711470 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1","Type":"ContainerDied","Data":"75456b9371694d2a2bfac3e7ae5718c1db9217ccac7f25b410e771f245a913db"} Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.711500 4744 scope.go:117] "RemoveContainer" containerID="dc5bd6fe3e4d8a622919d7c79c75c558165e9a11d9cbc232d117f75604e180d3" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.711658 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.719879 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-8bj8c" event={"ID":"df7dff0d-99e0-4e98-b752-9096d24e314b","Type":"ContainerDied","Data":"6553ff43656eeed87600ec11e174aab60f5d90c23e020ab2782ca19348b05eff"} Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.719926 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6553ff43656eeed87600ec11e174aab60f5d90c23e020ab2782ca19348b05eff" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.720017 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-8bj8c" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.774028 4744 scope.go:117] "RemoveContainer" containerID="0716d450e559e3ef093de38e99e9960e41173df8ae2deba8354f830a1a1c3623" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.774544 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.780130 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"fd08b448-b737-4aa8-b263-b48e3c32f03b","Type":"ContainerDied","Data":"7b5ce915256fb24bf7666c0205a50044a4691224696a890c8c3107a5eac8c614"} Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.808006 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.823035 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.837325 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.853184 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.908186 4744 scope.go:117] "RemoveContainer" containerID="dc5bd6fe3e4d8a622919d7c79c75c558165e9a11d9cbc232d117f75604e180d3" Oct 03 16:44:42 crc kubenswrapper[4744]: E1003 16:44:42.911373 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc5bd6fe3e4d8a622919d7c79c75c558165e9a11d9cbc232d117f75604e180d3\": container with ID starting with dc5bd6fe3e4d8a622919d7c79c75c558165e9a11d9cbc232d117f75604e180d3 not found: ID does not exist" containerID="dc5bd6fe3e4d8a622919d7c79c75c558165e9a11d9cbc232d117f75604e180d3" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.911589 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc5bd6fe3e4d8a622919d7c79c75c558165e9a11d9cbc232d117f75604e180d3"} err="failed to get container status \"dc5bd6fe3e4d8a622919d7c79c75c558165e9a11d9cbc232d117f75604e180d3\": rpc error: code = NotFound desc = could not find container \"dc5bd6fe3e4d8a622919d7c79c75c558165e9a11d9cbc232d117f75604e180d3\": container with ID starting with dc5bd6fe3e4d8a622919d7c79c75c558165e9a11d9cbc232d117f75604e180d3 not found: ID does not exist" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.911674 4744 scope.go:117] "RemoveContainer" containerID="0716d450e559e3ef093de38e99e9960e41173df8ae2deba8354f830a1a1c3623" Oct 03 16:44:42 crc kubenswrapper[4744]: E1003 16:44:42.914276 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0716d450e559e3ef093de38e99e9960e41173df8ae2deba8354f830a1a1c3623\": container with ID starting with 0716d450e559e3ef093de38e99e9960e41173df8ae2deba8354f830a1a1c3623 not found: ID does not exist" containerID="0716d450e559e3ef093de38e99e9960e41173df8ae2deba8354f830a1a1c3623" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.914418 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0716d450e559e3ef093de38e99e9960e41173df8ae2deba8354f830a1a1c3623"} err="failed to get container status \"0716d450e559e3ef093de38e99e9960e41173df8ae2deba8354f830a1a1c3623\": rpc error: code = NotFound desc = could not find container \"0716d450e559e3ef093de38e99e9960e41173df8ae2deba8354f830a1a1c3623\": container with ID starting with 0716d450e559e3ef093de38e99e9960e41173df8ae2deba8354f830a1a1c3623 not found: ID does not exist" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.914552 4744 scope.go:117] "RemoveContainer" containerID="7fb3b4a805e6dee19865f6ab7fd633ae4e6488f093a128f82810c7818bb044ea" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.916661 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21f7869b-e2f2-4e15-861c-9c3df31aa03e" path="/var/lib/kubelet/pods/21f7869b-e2f2-4e15-861c-9c3df31aa03e/volumes" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.918691 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.931150 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5171a06a-cac6-4fa5-abfc-8423d21003ce-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.942676 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-scripts\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.942748 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjcvh\" (UniqueName: \"kubernetes.io/projected/5171a06a-cac6-4fa5-abfc-8423d21003ce-kube-api-access-mjcvh\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.942774 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.942834 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-config-data\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.942927 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.961681 4744 scope.go:117] "RemoveContainer" containerID="359abd73c3de7bd2a15ad779b1c60725a6a8525c9fb24fc7a4089b2c6400b9e0" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.964632 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5865f9d689-cwtl8"] Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.969104 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:42 crc kubenswrapper[4744]: I1003 16:44:42.980058 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5865f9d689-cwtl8"] Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.006522 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.036559 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.046473 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5171a06a-cac6-4fa5-abfc-8423d21003ce-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.046549 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.046606 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-scripts\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.046638 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjcvh\" (UniqueName: \"kubernetes.io/projected/5171a06a-cac6-4fa5-abfc-8423d21003ce-kube-api-access-mjcvh\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.046657 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.046685 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/4fa451ca-3be8-4074-b903-5ea114f47bda-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.046702 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ld92\" (UniqueName: \"kubernetes.io/projected/4fa451ca-3be8-4074-b903-5ea114f47bda-kube-api-access-9ld92\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.046720 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4fa451ca-3be8-4074-b903-5ea114f47bda-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.046736 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-config-data\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.046753 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-config-data\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.046785 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-scripts\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.046815 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.046842 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4fa451ca-3be8-4074-b903-5ea114f47bda-ceph\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.046884 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.047151 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5171a06a-cac6-4fa5-abfc-8423d21003ce-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.052189 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.056925 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-scripts\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.057101 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.058729 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.059330 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-config-data\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.075187 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.078886 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.080234 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.089138 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjcvh\" (UniqueName: \"kubernetes.io/projected/5171a06a-cac6-4fa5-abfc-8423d21003ce-kube-api-access-mjcvh\") pod \"manila-scheduler-0\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " pod="openstack/manila-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.094588 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.113849 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.127920 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.129363 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.133846 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.138570 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.151589 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.153236 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.161762 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.161805 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4fa451ca-3be8-4074-b903-5ea114f47bda-ceph\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.161826 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-config\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.161843 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-ovsdbserver-sb\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.161864 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-config-data-custom\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.161887 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/146d5ce6-9a87-450e-b56c-371f2aa7af06-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.161903 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvl8d\" (UniqueName: \"kubernetes.io/projected/008783bb-cc87-416a-9823-787ff756a3c9-kube-api-access-nvl8d\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.161917 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008783bb-cc87-416a-9823-787ff756a3c9-scripts\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.161933 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.161953 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.161970 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.161993 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008783bb-cc87-416a-9823-787ff756a3c9-config-data\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162011 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162028 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-run\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162045 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/146d5ce6-9a87-450e-b56c-371f2aa7af06-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162059 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162078 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-ovsdbserver-nb\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162093 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-dns-swift-storage-0\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162113 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/146d5ce6-9a87-450e-b56c-371f2aa7af06-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162127 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-dns-svc\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162154 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-config-data\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162178 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-dev\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162192 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/008783bb-cc87-416a-9823-787ff756a3c9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162206 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msnkm\" (UniqueName: \"kubernetes.io/projected/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-kube-api-access-msnkm\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162219 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162235 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162250 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/146d5ce6-9a87-450e-b56c-371f2aa7af06-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162264 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162284 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-logs\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162305 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/146d5ce6-9a87-450e-b56c-371f2aa7af06-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162325 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/4fa451ca-3be8-4074-b903-5ea114f47bda-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162338 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ld92\" (UniqueName: \"kubernetes.io/projected/4fa451ca-3be8-4074-b903-5ea114f47bda-kube-api-access-9ld92\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162356 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4fa451ca-3be8-4074-b903-5ea114f47bda-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162370 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-config-data\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162387 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ln6h\" (UniqueName: \"kubernetes.io/projected/146d5ce6-9a87-450e-b56c-371f2aa7af06-kube-api-access-9ln6h\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162403 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-sys\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162421 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-scripts\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162435 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162451 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5qbd\" (UniqueName: \"kubernetes.io/projected/686a87c3-18f7-4578-a07c-ff9fb5a766cd-kube-api-access-g5qbd\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162472 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/008783bb-cc87-416a-9823-787ff756a3c9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162496 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-scripts\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162527 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008783bb-cc87-416a-9823-787ff756a3c9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.162547 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-etc-machine-id\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.174905 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.175295 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4fa451ca-3be8-4074-b903-5ea114f47bda-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.175561 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/4fa451ca-3be8-4074-b903-5ea114f47bda-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.179140 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.180262 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5467484d76-wdx4m" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.181739 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.187198 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-config-data\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.188495 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4fa451ca-3be8-4074-b903-5ea114f47bda-ceph\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.193318 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-scripts\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.194648 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.231300 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ld92\" (UniqueName: \"kubernetes.io/projected/4fa451ca-3be8-4074-b903-5ea114f47bda-kube-api-access-9ld92\") pod \"manila-share-share1-0\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.266590 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-config-data\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.266851 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-dev\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.266922 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msnkm\" (UniqueName: \"kubernetes.io/projected/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-kube-api-access-msnkm\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.266987 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.267093 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/008783bb-cc87-416a-9823-787ff756a3c9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.267161 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/146d5ce6-9a87-450e-b56c-371f2aa7af06-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.267224 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.267284 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.267352 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-logs\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.267431 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/146d5ce6-9a87-450e-b56c-371f2aa7af06-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.267525 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-sys\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.267587 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ln6h\" (UniqueName: \"kubernetes.io/projected/146d5ce6-9a87-450e-b56c-371f2aa7af06-kube-api-access-9ln6h\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.267663 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-scripts\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.267731 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.267805 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5qbd\" (UniqueName: \"kubernetes.io/projected/686a87c3-18f7-4578-a07c-ff9fb5a766cd-kube-api-access-g5qbd\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.267877 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/008783bb-cc87-416a-9823-787ff756a3c9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.268172 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008783bb-cc87-416a-9823-787ff756a3c9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.268252 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-etc-machine-id\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.268325 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.268405 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-config\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.268471 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-ovsdbserver-sb\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.268562 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-config-data-custom\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.268644 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/146d5ce6-9a87-450e-b56c-371f2aa7af06-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.268708 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvl8d\" (UniqueName: \"kubernetes.io/projected/008783bb-cc87-416a-9823-787ff756a3c9-kube-api-access-nvl8d\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.268784 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008783bb-cc87-416a-9823-787ff756a3c9-scripts\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.268852 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.268927 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.269005 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008783bb-cc87-416a-9823-787ff756a3c9-config-data\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.269073 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-run\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.269143 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/146d5ce6-9a87-450e-b56c-371f2aa7af06-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.269201 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-logs\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.269211 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.269291 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-ovsdbserver-nb\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.269319 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-dns-swift-storage-0\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.269362 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/146d5ce6-9a87-450e-b56c-371f2aa7af06-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.269379 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-dns-svc\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.269546 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.270258 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-dns-svc\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.270656 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.270692 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-dev\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.268073 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.273663 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.273740 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/008783bb-cc87-416a-9823-787ff756a3c9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.276955 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-sys\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.278533 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-etc-machine-id\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.278551 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-dns-swift-storage-0\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.278736 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.279028 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-ovsdbserver-nb\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.279398 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-config\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.279492 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/008783bb-cc87-416a-9823-787ff756a3c9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.279635 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.279897 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/146d5ce6-9a87-450e-b56c-371f2aa7af06-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.279981 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-run\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.280246 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008783bb-cc87-416a-9823-787ff756a3c9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.280377 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-ovsdbserver-sb\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.280980 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/146d5ce6-9a87-450e-b56c-371f2aa7af06-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.281062 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/146d5ce6-9a87-450e-b56c-371f2aa7af06-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.281291 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-scripts\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.285108 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/146d5ce6-9a87-450e-b56c-371f2aa7af06-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.286003 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008783bb-cc87-416a-9823-787ff756a3c9-scripts\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.286080 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-config-data\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.287154 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/146d5ce6-9a87-450e-b56c-371f2aa7af06-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.287640 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008783bb-cc87-416a-9823-787ff756a3c9-config-data\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.288101 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.290270 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-config-data-custom\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.290693 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/146d5ce6-9a87-450e-b56c-371f2aa7af06-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.299559 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msnkm\" (UniqueName: \"kubernetes.io/projected/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-kube-api-access-msnkm\") pod \"manila-api-0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.300274 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ln6h\" (UniqueName: \"kubernetes.io/projected/146d5ce6-9a87-450e-b56c-371f2aa7af06-kube-api-access-9ln6h\") pod \"cinder-volume-volume1-0\" (UID: \"146d5ce6-9a87-450e-b56c-371f2aa7af06\") " pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.303594 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5qbd\" (UniqueName: \"kubernetes.io/projected/686a87c3-18f7-4578-a07c-ff9fb5a766cd-kube-api-access-g5qbd\") pod \"dnsmasq-dns-5865f9d689-cwtl8\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.304017 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.307597 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.317027 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvl8d\" (UniqueName: \"kubernetes.io/projected/008783bb-cc87-416a-9823-787ff756a3c9-kube-api-access-nvl8d\") pod \"cinder-scheduler-0\" (UID: \"008783bb-cc87-416a-9823-787ff756a3c9\") " pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.317550 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.317984 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.509048 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.580711 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5467484d76-wdx4m" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:42312->10.217.0.162:9311: read: connection reset by peer" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.580979 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5467484d76-wdx4m" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:50424->10.217.0.162:9311: read: connection reset by peer" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.587894 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.797219 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.797810 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="ceilometer-central-agent" containerID="cri-o://64e1d8fa67bd4b27a329c86f45dd74b6928ba9f3ae1ce41f3a1deadae7090fa2" gracePeriod=30 Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.798182 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="proxy-httpd" containerID="cri-o://ba6406170f7106c9dba94384d81f63d45eb970af3d62d53c86e31b1dc85d47ff" gracePeriod=30 Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.798230 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="sg-core" containerID="cri-o://dc43d1e8772042d61a68de1f2b1c1c2ab881e61eac2e180894cd70a3c7153adf" gracePeriod=30 Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.798257 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="ceilometer-notification-agent" containerID="cri-o://a715f005eeddefc330647709836f3e4428fd6bf79cdba4079f4dda6559cfcf68" gracePeriod=30 Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.814290 4744 generic.go:334] "Generic (PLEG): container finished" podID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerID="96bf8b18b0a9bf7861ee3471bb012ef363c641506d0431562a761c26266e3de2" exitCode=0 Oct 03 16:44:43 crc kubenswrapper[4744]: I1003 16:44:43.814327 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5467484d76-wdx4m" event={"ID":"102bbbfd-8dd8-4946-9f35-a22a87ebaaed","Type":"ContainerDied","Data":"96bf8b18b0a9bf7861ee3471bb012ef363c641506d0431562a761c26266e3de2"} Oct 03 16:44:43 crc kubenswrapper[4744]: E1003 16:44:43.970946 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod102bbbfd_8dd8_4946_9f35_a22a87ebaaed.slice/crio-conmon-96bf8b18b0a9bf7861ee3471bb012ef363c641506d0431562a761c26266e3de2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod102bbbfd_8dd8_4946_9f35_a22a87ebaaed.slice/crio-96bf8b18b0a9bf7861ee3471bb012ef363c641506d0431562a761c26266e3de2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68631662_30f7_459c_bacc_8d326f8b8416.slice/crio-conmon-dc43d1e8772042d61a68de1f2b1c1c2ab881e61eac2e180894cd70a3c7153adf.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68631662_30f7_459c_bacc_8d326f8b8416.slice/crio-dc43d1e8772042d61a68de1f2b1c1c2ab881e61eac2e180894cd70a3c7153adf.scope\": RecentStats: unable to find data in memory cache]" Oct 03 16:44:44 crc kubenswrapper[4744]: W1003 16:44:44.014487 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod686a87c3_18f7_4578_a07c_ff9fb5a766cd.slice/crio-2293af597a99ea310d60af102b24a55a0d7fe2e8d4568c1d695276438bfdab28 WatchSource:0}: Error finding container 2293af597a99ea310d60af102b24a55a0d7fe2e8d4568c1d695276438bfdab28: Status 404 returned error can't find the container with id 2293af597a99ea310d60af102b24a55a0d7fe2e8d4568c1d695276438bfdab28 Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.023862 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.053011 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5865f9d689-cwtl8"] Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.094460 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.229666 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-backup-0"] Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.305483 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.321732 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:44 crc kubenswrapper[4744]: W1003 16:44:44.385237 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29a32f53_8017_4b24_ad95_d20e7ec1aaa0.slice/crio-974da30332c6d8e4475599f89ca0b8b3f869b05bc210e5b44c88bb2a740679b2 WatchSource:0}: Error finding container 974da30332c6d8e4475599f89ca0b8b3f869b05bc210e5b44c88bb2a740679b2: Status 404 returned error can't find the container with id 974da30332c6d8e4475599f89ca0b8b3f869b05bc210e5b44c88bb2a740679b2 Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.461626 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b2ks\" (UniqueName: \"kubernetes.io/projected/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-kube-api-access-6b2ks\") pod \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.461894 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-config-data\") pod \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.461921 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-combined-ca-bundle\") pod \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.462038 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-logs\") pod \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.462087 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-config-data-custom\") pod \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\" (UID: \"102bbbfd-8dd8-4946-9f35-a22a87ebaaed\") " Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.465482 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.466834 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "102bbbfd-8dd8-4946-9f35-a22a87ebaaed" (UID: "102bbbfd-8dd8-4946-9f35-a22a87ebaaed"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.471697 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-logs" (OuterVolumeSpecName: "logs") pod "102bbbfd-8dd8-4946-9f35-a22a87ebaaed" (UID: "102bbbfd-8dd8-4946-9f35-a22a87ebaaed"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.475207 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-kube-api-access-6b2ks" (OuterVolumeSpecName: "kube-api-access-6b2ks") pod "102bbbfd-8dd8-4946-9f35-a22a87ebaaed" (UID: "102bbbfd-8dd8-4946-9f35-a22a87ebaaed"). InnerVolumeSpecName "kube-api-access-6b2ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.531925 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "102bbbfd-8dd8-4946-9f35-a22a87ebaaed" (UID: "102bbbfd-8dd8-4946-9f35-a22a87ebaaed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.534460 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.541451 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-config-data" (OuterVolumeSpecName: "config-data") pod "102bbbfd-8dd8-4946-9f35-a22a87ebaaed" (UID: "102bbbfd-8dd8-4946-9f35-a22a87ebaaed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.564182 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b2ks\" (UniqueName: \"kubernetes.io/projected/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-kube-api-access-6b2ks\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.564221 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.564232 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.564241 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.564251 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/102bbbfd-8dd8-4946-9f35-a22a87ebaaed-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.577857 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 03 16:44:44 crc kubenswrapper[4744]: W1003 16:44:44.641295 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008783bb_cc87_416a_9823_787ff756a3c9.slice/crio-b836b7505dfb2d57e9a0936f967934423a2fc02d7a331849ea52c4353ac24508 WatchSource:0}: Error finding container b836b7505dfb2d57e9a0936f967934423a2fc02d7a331849ea52c4353ac24508: Status 404 returned error can't find the container with id b836b7505dfb2d57e9a0936f967934423a2fc02d7a331849ea52c4353ac24508 Oct 03 16:44:44 crc kubenswrapper[4744]: W1003 16:44:44.648425 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod146d5ce6_9a87_450e_b56c_371f2aa7af06.slice/crio-5c3ce18fc2352f0c475cbaa8733e862d797865d1d4c6575e790a0735d1894139 WatchSource:0}: Error finding container 5c3ce18fc2352f0c475cbaa8733e862d797865d1d4c6575e790a0735d1894139: Status 404 returned error can't find the container with id 5c3ce18fc2352f0c475cbaa8733e862d797865d1d4c6575e790a0735d1894139 Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.827338 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"29a32f53-8017-4b24-ad95-d20e7ec1aaa0","Type":"ContainerStarted","Data":"974da30332c6d8e4475599f89ca0b8b3f869b05bc210e5b44c88bb2a740679b2"} Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.830027 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"146d5ce6-9a87-450e-b56c-371f2aa7af06","Type":"ContainerStarted","Data":"5c3ce18fc2352f0c475cbaa8733e862d797865d1d4c6575e790a0735d1894139"} Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.832961 4744 generic.go:334] "Generic (PLEG): container finished" podID="686a87c3-18f7-4578-a07c-ff9fb5a766cd" containerID="f280fc967e97cb2fff2a4a3b560febe1c9455a1f3ff75052ded866b0a151aee6" exitCode=0 Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.833089 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" event={"ID":"686a87c3-18f7-4578-a07c-ff9fb5a766cd","Type":"ContainerDied","Data":"f280fc967e97cb2fff2a4a3b560febe1c9455a1f3ff75052ded866b0a151aee6"} Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.833123 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" event={"ID":"686a87c3-18f7-4578-a07c-ff9fb5a766cd","Type":"ContainerStarted","Data":"2293af597a99ea310d60af102b24a55a0d7fe2e8d4568c1d695276438bfdab28"} Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.840195 4744 generic.go:334] "Generic (PLEG): container finished" podID="68631662-30f7-459c-bacc-8d326f8b8416" containerID="ba6406170f7106c9dba94384d81f63d45eb970af3d62d53c86e31b1dc85d47ff" exitCode=0 Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.840220 4744 generic.go:334] "Generic (PLEG): container finished" podID="68631662-30f7-459c-bacc-8d326f8b8416" containerID="dc43d1e8772042d61a68de1f2b1c1c2ab881e61eac2e180894cd70a3c7153adf" exitCode=2 Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.840229 4744 generic.go:334] "Generic (PLEG): container finished" podID="68631662-30f7-459c-bacc-8d326f8b8416" containerID="a715f005eeddefc330647709836f3e4428fd6bf79cdba4079f4dda6559cfcf68" exitCode=0 Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.840235 4744 generic.go:334] "Generic (PLEG): container finished" podID="68631662-30f7-459c-bacc-8d326f8b8416" containerID="64e1d8fa67bd4b27a329c86f45dd74b6928ba9f3ae1ce41f3a1deadae7090fa2" exitCode=0 Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.840276 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68631662-30f7-459c-bacc-8d326f8b8416","Type":"ContainerDied","Data":"ba6406170f7106c9dba94384d81f63d45eb970af3d62d53c86e31b1dc85d47ff"} Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.840300 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68631662-30f7-459c-bacc-8d326f8b8416","Type":"ContainerDied","Data":"dc43d1e8772042d61a68de1f2b1c1c2ab881e61eac2e180894cd70a3c7153adf"} Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.840311 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68631662-30f7-459c-bacc-8d326f8b8416","Type":"ContainerDied","Data":"a715f005eeddefc330647709836f3e4428fd6bf79cdba4079f4dda6559cfcf68"} Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.840319 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68631662-30f7-459c-bacc-8d326f8b8416","Type":"ContainerDied","Data":"64e1d8fa67bd4b27a329c86f45dd74b6928ba9f3ae1ce41f3a1deadae7090fa2"} Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.840327 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"68631662-30f7-459c-bacc-8d326f8b8416","Type":"ContainerDied","Data":"5a6ee26bfe526aa0d8dedeffe84b22d74f74207896a02f59d5ec44f4a94e362b"} Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.840336 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a6ee26bfe526aa0d8dedeffe84b22d74f74207896a02f59d5ec44f4a94e362b" Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.844415 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5171a06a-cac6-4fa5-abfc-8423d21003ce","Type":"ContainerStarted","Data":"a62d6b537ef68100ec667232e02233821286e60b44bc353ee01fd24d8a68f4bc"} Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.847610 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"4fa451ca-3be8-4074-b903-5ea114f47bda","Type":"ContainerStarted","Data":"fa4c3c41d905a5c33b1f8bc53a6bab59c974a52e53b72f7b6ffface0dcbd30d0"} Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.851698 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"008783bb-cc87-416a-9823-787ff756a3c9","Type":"ContainerStarted","Data":"b836b7505dfb2d57e9a0936f967934423a2fc02d7a331849ea52c4353ac24508"} Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.875122 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-backup-0" podUID="65ff9f68-860d-41e6-a48b-4ada29ea6275" containerName="cinder-backup" containerID="cri-o://b61091089b1f64d9c36837e9b95c7f9363f084dd5fa2df3f9e7e99746f932811" gracePeriod=30 Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.875250 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.875323 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5467484d76-wdx4m" event={"ID":"102bbbfd-8dd8-4946-9f35-a22a87ebaaed","Type":"ContainerDied","Data":"77d1c3e3a61cfe2d3ed30b7fb39f2af9e7f613712d7f3921d9fa9e1efdc7819c"} Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.875352 4744 scope.go:117] "RemoveContainer" containerID="96bf8b18b0a9bf7861ee3471bb012ef363c641506d0431562a761c26266e3de2" Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.875565 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-backup-0" podUID="65ff9f68-860d-41e6-a48b-4ada29ea6275" containerName="probe" containerID="cri-o://17e2d0409805a5ae4d40c9353c78ea0bda5417cc741323264f56f0eff8e26aca" gracePeriod=30 Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.971058 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1" path="/var/lib/kubelet/pods/4b8aaeb5-5dbd-4c09-8b25-dc00ed12c7c1/volumes" Oct 03 16:44:44 crc kubenswrapper[4744]: I1003 16:44:44.972090 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd08b448-b737-4aa8-b263-b48e3c32f03b" path="/var/lib/kubelet/pods/fd08b448-b737-4aa8-b263-b48e3c32f03b/volumes" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.075532 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.152426 4744 scope.go:117] "RemoveContainer" containerID="8a8ff0666730ce0e751b72083f365fc1952aa3b0d6aeba77566dba6aead408c9" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.194918 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-config-data\") pod \"68631662-30f7-459c-bacc-8d326f8b8416\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.195043 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-sg-core-conf-yaml\") pod \"68631662-30f7-459c-bacc-8d326f8b8416\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.195317 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-scripts\") pod \"68631662-30f7-459c-bacc-8d326f8b8416\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.195427 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68631662-30f7-459c-bacc-8d326f8b8416-log-httpd\") pod \"68631662-30f7-459c-bacc-8d326f8b8416\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.195467 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68631662-30f7-459c-bacc-8d326f8b8416-run-httpd\") pod \"68631662-30f7-459c-bacc-8d326f8b8416\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.195604 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-combined-ca-bundle\") pod \"68631662-30f7-459c-bacc-8d326f8b8416\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.195628 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbf5f\" (UniqueName: \"kubernetes.io/projected/68631662-30f7-459c-bacc-8d326f8b8416-kube-api-access-pbf5f\") pod \"68631662-30f7-459c-bacc-8d326f8b8416\" (UID: \"68631662-30f7-459c-bacc-8d326f8b8416\") " Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.202929 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-scripts" (OuterVolumeSpecName: "scripts") pod "68631662-30f7-459c-bacc-8d326f8b8416" (UID: "68631662-30f7-459c-bacc-8d326f8b8416"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.208523 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68631662-30f7-459c-bacc-8d326f8b8416-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "68631662-30f7-459c-bacc-8d326f8b8416" (UID: "68631662-30f7-459c-bacc-8d326f8b8416"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.212788 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68631662-30f7-459c-bacc-8d326f8b8416-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "68631662-30f7-459c-bacc-8d326f8b8416" (UID: "68631662-30f7-459c-bacc-8d326f8b8416"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.250616 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68631662-30f7-459c-bacc-8d326f8b8416-kube-api-access-pbf5f" (OuterVolumeSpecName: "kube-api-access-pbf5f") pod "68631662-30f7-459c-bacc-8d326f8b8416" (UID: "68631662-30f7-459c-bacc-8d326f8b8416"). InnerVolumeSpecName "kube-api-access-pbf5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.297865 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbf5f\" (UniqueName: \"kubernetes.io/projected/68631662-30f7-459c-bacc-8d326f8b8416-kube-api-access-pbf5f\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.298196 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.298283 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68631662-30f7-459c-bacc-8d326f8b8416-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.298351 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68631662-30f7-459c-bacc-8d326f8b8416-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.300683 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "68631662-30f7-459c-bacc-8d326f8b8416" (UID: "68631662-30f7-459c-bacc-8d326f8b8416"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.401267 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.567637 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68631662-30f7-459c-bacc-8d326f8b8416" (UID: "68631662-30f7-459c-bacc-8d326f8b8416"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.607541 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.740861 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-config-data" (OuterVolumeSpecName: "config-data") pod "68631662-30f7-459c-bacc-8d326f8b8416" (UID: "68631662-30f7-459c-bacc-8d326f8b8416"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.820000 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68631662-30f7-459c-bacc-8d326f8b8416-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.914191 4744 generic.go:334] "Generic (PLEG): container finished" podID="65ff9f68-860d-41e6-a48b-4ada29ea6275" containerID="17e2d0409805a5ae4d40c9353c78ea0bda5417cc741323264f56f0eff8e26aca" exitCode=0 Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.914986 4744 generic.go:334] "Generic (PLEG): container finished" podID="65ff9f68-860d-41e6-a48b-4ada29ea6275" containerID="b61091089b1f64d9c36837e9b95c7f9363f084dd5fa2df3f9e7e99746f932811" exitCode=0 Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.915038 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"65ff9f68-860d-41e6-a48b-4ada29ea6275","Type":"ContainerDied","Data":"17e2d0409805a5ae4d40c9353c78ea0bda5417cc741323264f56f0eff8e26aca"} Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.915064 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"65ff9f68-860d-41e6-a48b-4ada29ea6275","Type":"ContainerDied","Data":"b61091089b1f64d9c36837e9b95c7f9363f084dd5fa2df3f9e7e99746f932811"} Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.944753 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"008783bb-cc87-416a-9823-787ff756a3c9","Type":"ContainerStarted","Data":"7255a06d3b387ce678b749e4321af8620a624a2c15ace419ff2a7ca174c20533"} Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.995282 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"146d5ce6-9a87-450e-b56c-371f2aa7af06","Type":"ContainerStarted","Data":"65c8557573b3656947fc00b2e8ef1867a6b820f31379b82a1d7b46fdb5dad279"} Oct 03 16:44:45 crc kubenswrapper[4744]: I1003 16:44:45.995585 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"146d5ce6-9a87-450e-b56c-371f2aa7af06","Type":"ContainerStarted","Data":"01cf1e81c9df2105317f314f1cd53959c5373b4ace17ab11bc40e7f53840d183"} Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.069257 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"29a32f53-8017-4b24-ad95-d20e7ec1aaa0","Type":"ContainerStarted","Data":"075ed500d2422c1e0176fa45154326ae243f99953a80fd5d7e971650fedaad44"} Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.142165 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" event={"ID":"686a87c3-18f7-4578-a07c-ff9fb5a766cd","Type":"ContainerStarted","Data":"53872c77e10672dc21a6ea541bf521cc51f9325fd6745aca8761d4721a9c4097"} Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.143236 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.177885 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.194723 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=4.194699466 podStartE2EDuration="4.194699466s" podCreationTimestamp="2025-10-03 16:44:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:46.068079166 +0000 UTC m=+1212.347955062" watchObservedRunningTime="2025-10-03 16:44:46.194699466 +0000 UTC m=+1212.474575362" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.196671 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.220393 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-machine-id\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.220726 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-sys\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.220795 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" podStartSLOduration=4.220768047 podStartE2EDuration="4.220768047s" podCreationTimestamp="2025-10-03 16:44:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:46.189423336 +0000 UTC m=+1212.469299232" watchObservedRunningTime="2025-10-03 16:44:46.220768047 +0000 UTC m=+1212.500643953" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.220593 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.222240 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-sys" (OuterVolumeSpecName: "sys") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.220847 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-combined-ca-bundle\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.222958 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-nvme\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.222994 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-run\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223063 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223088 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-run" (OuterVolumeSpecName: "run") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223105 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-config-data\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223132 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-lib-cinder\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223399 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-dev\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223416 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-locks-brick\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223435 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-lib-modules\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223507 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223552 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-dev" (OuterVolumeSpecName: "dev") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223601 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223574 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223567 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lj4sb\" (UniqueName: \"kubernetes.io/projected/65ff9f68-860d-41e6-a48b-4ada29ea6275-kube-api-access-lj4sb\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223747 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-iscsi\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223798 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-config-data-custom\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223833 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/65ff9f68-860d-41e6-a48b-4ada29ea6275-ceph\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223898 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-locks-cinder\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.223943 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-scripts\") pod \"65ff9f68-860d-41e6-a48b-4ada29ea6275\" (UID: \"65ff9f68-860d-41e6-a48b-4ada29ea6275\") " Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.225487 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.225635 4744 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-dev\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.225667 4744 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.225696 4744 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.225708 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.225722 4744 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-sys\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.225732 4744 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.225744 4744 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-run\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.225754 4744 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.226660 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.226703 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.232719 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.238031 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-scripts" (OuterVolumeSpecName: "scripts") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.264816 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ff9f68-860d-41e6-a48b-4ada29ea6275-ceph" (OuterVolumeSpecName: "ceph") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.272050 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.273137 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ff9f68-860d-41e6-a48b-4ada29ea6275-kube-api-access-lj4sb" (OuterVolumeSpecName: "kube-api-access-lj4sb") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "kube-api-access-lj4sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.280843 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:46 crc kubenswrapper[4744]: E1003 16:44:46.281353 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ff9f68-860d-41e6-a48b-4ada29ea6275" containerName="probe" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.281378 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ff9f68-860d-41e6-a48b-4ada29ea6275" containerName="probe" Oct 03 16:44:46 crc kubenswrapper[4744]: E1003 16:44:46.281405 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.281415 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api" Oct 03 16:44:46 crc kubenswrapper[4744]: E1003 16:44:46.281435 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="ceilometer-central-agent" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.281442 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="ceilometer-central-agent" Oct 03 16:44:46 crc kubenswrapper[4744]: E1003 16:44:46.281464 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api-log" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.281476 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api-log" Oct 03 16:44:46 crc kubenswrapper[4744]: E1003 16:44:46.281489 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ff9f68-860d-41e6-a48b-4ada29ea6275" containerName="cinder-backup" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.281499 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ff9f68-860d-41e6-a48b-4ada29ea6275" containerName="cinder-backup" Oct 03 16:44:46 crc kubenswrapper[4744]: E1003 16:44:46.281678 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="proxy-httpd" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.281689 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="proxy-httpd" Oct 03 16:44:46 crc kubenswrapper[4744]: E1003 16:44:46.281710 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="ceilometer-notification-agent" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.281716 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="ceilometer-notification-agent" Oct 03 16:44:46 crc kubenswrapper[4744]: E1003 16:44:46.281729 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="sg-core" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.281738 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="sg-core" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.281958 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="65ff9f68-860d-41e6-a48b-4ada29ea6275" containerName="cinder-backup" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.281984 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="ceilometer-notification-agent" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.281996 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api-log" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.282011 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="ceilometer-central-agent" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.282022 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.282034 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="65ff9f68-860d-41e6-a48b-4ada29ea6275" containerName="probe" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.282056 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="sg-core" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.282067 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="68631662-30f7-459c-bacc-8d326f8b8416" containerName="proxy-httpd" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.289146 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.297475 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.297960 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.313662 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.328132 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m72pc\" (UniqueName: \"kubernetes.io/projected/34b80b77-d825-45d7-825c-88a8089fa4ca-kube-api-access-m72pc\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.328241 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34b80b77-d825-45d7-825c-88a8089fa4ca-run-httpd\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.328304 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-config-data\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.328333 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34b80b77-d825-45d7-825c-88a8089fa4ca-log-httpd\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.328435 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-scripts\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.328471 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.328563 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.328671 4744 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.328691 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.328705 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/65ff9f68-860d-41e6-a48b-4ada29ea6275-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.328717 4744 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/65ff9f68-860d-41e6-a48b-4ada29ea6275-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.328729 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.328741 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lj4sb\" (UniqueName: \"kubernetes.io/projected/65ff9f68-860d-41e6-a48b-4ada29ea6275-kube-api-access-lj4sb\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.431470 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34b80b77-d825-45d7-825c-88a8089fa4ca-run-httpd\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.431994 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34b80b77-d825-45d7-825c-88a8089fa4ca-run-httpd\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.431834 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-config-data\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.434384 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34b80b77-d825-45d7-825c-88a8089fa4ca-log-httpd\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.434590 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34b80b77-d825-45d7-825c-88a8089fa4ca-log-httpd\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.434817 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-scripts\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.434862 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.434902 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.435030 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m72pc\" (UniqueName: \"kubernetes.io/projected/34b80b77-d825-45d7-825c-88a8089fa4ca-kube-api-access-m72pc\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.439598 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.441047 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-config-data\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.441900 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.444312 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-scripts\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.458619 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-config-data" (OuterVolumeSpecName: "config-data") pod "65ff9f68-860d-41e6-a48b-4ada29ea6275" (UID: "65ff9f68-860d-41e6-a48b-4ada29ea6275"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.459276 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.464777 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m72pc\" (UniqueName: \"kubernetes.io/projected/34b80b77-d825-45d7-825c-88a8089fa4ca-kube-api-access-m72pc\") pod \"ceilometer-0\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.469321 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.537945 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.537980 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65ff9f68-860d-41e6-a48b-4ada29ea6275-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.643872 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:44:46 crc kubenswrapper[4744]: I1003 16:44:46.991181 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68631662-30f7-459c-bacc-8d326f8b8416" path="/var/lib/kubelet/pods/68631662-30f7-459c-bacc-8d326f8b8416/volumes" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.234688 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5171a06a-cac6-4fa5-abfc-8423d21003ce","Type":"ContainerStarted","Data":"e7e73d5de0975ca0698a8f0cd60545f25b1cfea11b338e43e683a3199a0a9036"} Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.235005 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5171a06a-cac6-4fa5-abfc-8423d21003ce","Type":"ContainerStarted","Data":"39ccc5d10f57ae51069eed00f8c31eb82d8557f26f7e79032a711276b096ef2a"} Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.260812 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=4.610704887 podStartE2EDuration="5.260798314s" podCreationTimestamp="2025-10-03 16:44:42 +0000 UTC" firstStartedPulling="2025-10-03 16:44:44.02888955 +0000 UTC m=+1210.308765436" lastFinishedPulling="2025-10-03 16:44:44.678982967 +0000 UTC m=+1210.958858863" observedRunningTime="2025-10-03 16:44:47.258727427 +0000 UTC m=+1213.538603323" watchObservedRunningTime="2025-10-03 16:44:47.260798314 +0000 UTC m=+1213.540674210" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.267744 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"65ff9f68-860d-41e6-a48b-4ada29ea6275","Type":"ContainerDied","Data":"5c3820a2d701f7dc7a943209600289be7de2c96f9df9917f39bd3aa5e4ed0347"} Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.267793 4744 scope.go:117] "RemoveContainer" containerID="17e2d0409805a5ae4d40c9353c78ea0bda5417cc741323264f56f0eff8e26aca" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.267910 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.298736 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"008783bb-cc87-416a-9823-787ff756a3c9","Type":"ContainerStarted","Data":"f06e12036a4c57f7ca361c345dae2319a76341da59b7ac8dcc7e9eece090b324"} Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.302908 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-backup-0"] Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.312394 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="29a32f53-8017-4b24-ad95-d20e7ec1aaa0" containerName="manila-api-log" containerID="cri-o://075ed500d2422c1e0176fa45154326ae243f99953a80fd5d7e971650fedaad44" gracePeriod=30 Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.312695 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="29a32f53-8017-4b24-ad95-d20e7ec1aaa0" containerName="manila-api" containerID="cri-o://4517f45c77679c3f0d853c515823705bf5abb55c901b955cc81e2fd121d4357a" gracePeriod=30 Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.312717 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"29a32f53-8017-4b24-ad95-d20e7ec1aaa0","Type":"ContainerStarted","Data":"4517f45c77679c3f0d853c515823705bf5abb55c901b955cc81e2fd121d4357a"} Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.312755 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.333791 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-backup-0"] Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.385752 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.386203 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.386186326 podStartE2EDuration="5.386186326s" podCreationTimestamp="2025-10-03 16:44:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:47.32592992 +0000 UTC m=+1213.605805806" watchObservedRunningTime="2025-10-03 16:44:47.386186326 +0000 UTC m=+1213.666062222" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.388067 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.391452 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.417147 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.431555 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.435273 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=5.435246548 podStartE2EDuration="5.435246548s" podCreationTimestamp="2025-10-03 16:44:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:47.36210587 +0000 UTC m=+1213.641981766" watchObservedRunningTime="2025-10-03 16:44:47.435246548 +0000 UTC m=+1213.715122444" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.503961 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.504037 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33874615-0e1f-4743-aa84-d262b6eab68c-config-data\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.504085 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.504116 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-lib-modules\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.504140 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33874615-0e1f-4743-aa84-d262b6eab68c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.504177 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98rpg\" (UniqueName: \"kubernetes.io/projected/33874615-0e1f-4743-aa84-d262b6eab68c-kube-api-access-98rpg\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.504213 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.504233 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33874615-0e1f-4743-aa84-d262b6eab68c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.504265 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33874615-0e1f-4743-aa84-d262b6eab68c-scripts\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.504295 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-dev\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.504322 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.504345 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/33874615-0e1f-4743-aa84-d262b6eab68c-ceph\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.504367 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.504398 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.504434 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-sys\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.504459 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-run\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.608829 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-sys\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.608875 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-run\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.608926 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.608961 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33874615-0e1f-4743-aa84-d262b6eab68c-config-data\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.608995 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.609019 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-lib-modules\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.609041 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33874615-0e1f-4743-aa84-d262b6eab68c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.609069 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98rpg\" (UniqueName: \"kubernetes.io/projected/33874615-0e1f-4743-aa84-d262b6eab68c-kube-api-access-98rpg\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.609094 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.609108 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33874615-0e1f-4743-aa84-d262b6eab68c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.609132 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33874615-0e1f-4743-aa84-d262b6eab68c-scripts\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.609157 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-dev\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.609175 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.609195 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/33874615-0e1f-4743-aa84-d262b6eab68c-ceph\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.609213 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.609238 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.609323 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.609359 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-sys\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.609377 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-run\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.609416 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.611407 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.611434 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.611512 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.612128 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-dev\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.612169 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.612214 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/33874615-0e1f-4743-aa84-d262b6eab68c-lib-modules\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.631014 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/33874615-0e1f-4743-aa84-d262b6eab68c-ceph\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.631250 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/33874615-0e1f-4743-aa84-d262b6eab68c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.634982 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33874615-0e1f-4743-aa84-d262b6eab68c-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.637088 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98rpg\" (UniqueName: \"kubernetes.io/projected/33874615-0e1f-4743-aa84-d262b6eab68c-kube-api-access-98rpg\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.645955 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33874615-0e1f-4743-aa84-d262b6eab68c-scripts\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.646061 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33874615-0e1f-4743-aa84-d262b6eab68c-config-data\") pod \"cinder-backup-0\" (UID: \"33874615-0e1f-4743-aa84-d262b6eab68c\") " pod="openstack/cinder-backup-0" Oct 03 16:44:47 crc kubenswrapper[4744]: I1003 16:44:47.716030 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 03 16:44:48 crc kubenswrapper[4744]: I1003 16:44:48.309296 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:48 crc kubenswrapper[4744]: I1003 16:44:48.329152 4744 generic.go:334] "Generic (PLEG): container finished" podID="29a32f53-8017-4b24-ad95-d20e7ec1aaa0" containerID="4517f45c77679c3f0d853c515823705bf5abb55c901b955cc81e2fd121d4357a" exitCode=0 Oct 03 16:44:48 crc kubenswrapper[4744]: I1003 16:44:48.329183 4744 generic.go:334] "Generic (PLEG): container finished" podID="29a32f53-8017-4b24-ad95-d20e7ec1aaa0" containerID="075ed500d2422c1e0176fa45154326ae243f99953a80fd5d7e971650fedaad44" exitCode=143 Oct 03 16:44:48 crc kubenswrapper[4744]: I1003 16:44:48.329330 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"29a32f53-8017-4b24-ad95-d20e7ec1aaa0","Type":"ContainerDied","Data":"4517f45c77679c3f0d853c515823705bf5abb55c901b955cc81e2fd121d4357a"} Oct 03 16:44:48 crc kubenswrapper[4744]: I1003 16:44:48.329375 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"29a32f53-8017-4b24-ad95-d20e7ec1aaa0","Type":"ContainerDied","Data":"075ed500d2422c1e0176fa45154326ae243f99953a80fd5d7e971650fedaad44"} Oct 03 16:44:48 crc kubenswrapper[4744]: I1003 16:44:48.589195 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 03 16:44:48 crc kubenswrapper[4744]: I1003 16:44:48.912306 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65ff9f68-860d-41e6-a48b-4ada29ea6275" path="/var/lib/kubelet/pods/65ff9f68-860d-41e6-a48b-4ada29ea6275/volumes" Oct 03 16:44:49 crc kubenswrapper[4744]: I1003 16:44:49.162273 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5467484d76-wdx4m" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 16:44:49 crc kubenswrapper[4744]: I1003 16:44:49.162279 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5467484d76-wdx4m" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 16:44:49 crc kubenswrapper[4744]: I1003 16:44:49.165066 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:44:49 crc kubenswrapper[4744]: I1003 16:44:49.685200 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:49 crc kubenswrapper[4744]: I1003 16:44:49.761932 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-774fb64f68-qgzkk" Oct 03 16:44:51 crc kubenswrapper[4744]: I1003 16:44:51.034350 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 03 16:44:53 crc kubenswrapper[4744]: I1003 16:44:53.307175 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:44:53 crc kubenswrapper[4744]: I1003 16:44:53.318412 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 03 16:44:53 crc kubenswrapper[4744]: I1003 16:44:53.399185 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qcrs4"] Oct 03 16:44:53 crc kubenswrapper[4744]: I1003 16:44:53.399414 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" podUID="4047d3e2-47ab-412e-a491-34835eac6716" containerName="dnsmasq-dns" containerID="cri-o://6ecd7f73349ba82f8168f463061e4ccaea47660f8e0a3feb08119ad164a7be4e" gracePeriod=10 Oct 03 16:44:53 crc kubenswrapper[4744]: I1003 16:44:53.573086 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 03 16:44:53 crc kubenswrapper[4744]: I1003 16:44:53.841463 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 03 16:44:54 crc kubenswrapper[4744]: I1003 16:44:54.353857 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" podUID="4047d3e2-47ab-412e-a491-34835eac6716" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.167:5353: connect: connection refused" Oct 03 16:44:54 crc kubenswrapper[4744]: I1003 16:44:54.429325 4744 generic.go:334] "Generic (PLEG): container finished" podID="4047d3e2-47ab-412e-a491-34835eac6716" containerID="6ecd7f73349ba82f8168f463061e4ccaea47660f8e0a3feb08119ad164a7be4e" exitCode=0 Oct 03 16:44:54 crc kubenswrapper[4744]: I1003 16:44:54.429369 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" event={"ID":"4047d3e2-47ab-412e-a491-34835eac6716","Type":"ContainerDied","Data":"6ecd7f73349ba82f8168f463061e4ccaea47660f8e0a3feb08119ad164a7be4e"} Oct 03 16:44:54 crc kubenswrapper[4744]: I1003 16:44:54.980629 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-wh96r"] Oct 03 16:44:54 crc kubenswrapper[4744]: I1003 16:44:54.982103 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wh96r" Oct 03 16:44:54 crc kubenswrapper[4744]: I1003 16:44:54.999442 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-wh96r"] Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.082608 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzl6s\" (UniqueName: \"kubernetes.io/projected/cca8b080-5315-4c0e-ab9d-c94c248df68d-kube-api-access-kzl6s\") pod \"nova-api-db-create-wh96r\" (UID: \"cca8b080-5315-4c0e-ab9d-c94c248df68d\") " pod="openstack/nova-api-db-create-wh96r" Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.160803 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-ldg2v"] Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.161977 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ldg2v" Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.180445 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-ldg2v"] Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.184103 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjq6l\" (UniqueName: \"kubernetes.io/projected/15b5b2d9-3e50-4326-a345-8b5f79d684e2-kube-api-access-bjq6l\") pod \"nova-cell0-db-create-ldg2v\" (UID: \"15b5b2d9-3e50-4326-a345-8b5f79d684e2\") " pod="openstack/nova-cell0-db-create-ldg2v" Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.184521 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzl6s\" (UniqueName: \"kubernetes.io/projected/cca8b080-5315-4c0e-ab9d-c94c248df68d-kube-api-access-kzl6s\") pod \"nova-api-db-create-wh96r\" (UID: \"cca8b080-5315-4c0e-ab9d-c94c248df68d\") " pod="openstack/nova-api-db-create-wh96r" Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.208363 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzl6s\" (UniqueName: \"kubernetes.io/projected/cca8b080-5315-4c0e-ab9d-c94c248df68d-kube-api-access-kzl6s\") pod \"nova-api-db-create-wh96r\" (UID: \"cca8b080-5315-4c0e-ab9d-c94c248df68d\") " pod="openstack/nova-api-db-create-wh96r" Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.271175 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-qsplt"] Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.272675 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qsplt" Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.281327 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-qsplt"] Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.286571 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsxv9\" (UniqueName: \"kubernetes.io/projected/501a4d70-6aa9-4fd5-8e72-5b854e648219-kube-api-access-gsxv9\") pod \"nova-cell1-db-create-qsplt\" (UID: \"501a4d70-6aa9-4fd5-8e72-5b854e648219\") " pod="openstack/nova-cell1-db-create-qsplt" Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.286644 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjq6l\" (UniqueName: \"kubernetes.io/projected/15b5b2d9-3e50-4326-a345-8b5f79d684e2-kube-api-access-bjq6l\") pod \"nova-cell0-db-create-ldg2v\" (UID: \"15b5b2d9-3e50-4326-a345-8b5f79d684e2\") " pod="openstack/nova-cell0-db-create-ldg2v" Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.316911 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjq6l\" (UniqueName: \"kubernetes.io/projected/15b5b2d9-3e50-4326-a345-8b5f79d684e2-kube-api-access-bjq6l\") pod \"nova-cell0-db-create-ldg2v\" (UID: \"15b5b2d9-3e50-4326-a345-8b5f79d684e2\") " pod="openstack/nova-cell0-db-create-ldg2v" Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.318874 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wh96r" Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.388528 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsxv9\" (UniqueName: \"kubernetes.io/projected/501a4d70-6aa9-4fd5-8e72-5b854e648219-kube-api-access-gsxv9\") pod \"nova-cell1-db-create-qsplt\" (UID: \"501a4d70-6aa9-4fd5-8e72-5b854e648219\") " pod="openstack/nova-cell1-db-create-qsplt" Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.448753 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsxv9\" (UniqueName: \"kubernetes.io/projected/501a4d70-6aa9-4fd5-8e72-5b854e648219-kube-api-access-gsxv9\") pod \"nova-cell1-db-create-qsplt\" (UID: \"501a4d70-6aa9-4fd5-8e72-5b854e648219\") " pod="openstack/nova-cell1-db-create-qsplt" Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.467908 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34b80b77-d825-45d7-825c-88a8089fa4ca","Type":"ContainerStarted","Data":"5d2258829ece1902f0ebc0685a6d055a1a6a3b65a93fe1af30b94ee9d4547789"} Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.492986 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ldg2v" Oct 03 16:44:55 crc kubenswrapper[4744]: I1003 16:44:55.679657 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qsplt" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.524860 4744 scope.go:117] "RemoveContainer" containerID="b61091089b1f64d9c36837e9b95c7f9363f084dd5fa2df3f9e7e99746f932811" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.777279 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.781352 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.784035 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-ovsdbserver-nb\") pod \"4047d3e2-47ab-412e-a491-34835eac6716\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.784109 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-config-data-custom\") pod \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.784128 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-dns-swift-storage-0\") pod \"4047d3e2-47ab-412e-a491-34835eac6716\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.784209 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-scripts\") pod \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.784239 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-dns-svc\") pod \"4047d3e2-47ab-412e-a491-34835eac6716\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.784256 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-ovsdbserver-sb\") pod \"4047d3e2-47ab-412e-a491-34835eac6716\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.784277 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppjtf\" (UniqueName: \"kubernetes.io/projected/4047d3e2-47ab-412e-a491-34835eac6716-kube-api-access-ppjtf\") pod \"4047d3e2-47ab-412e-a491-34835eac6716\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.784322 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-config-data\") pod \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.784377 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-etc-machine-id\") pod \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.784394 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-config\") pod \"4047d3e2-47ab-412e-a491-34835eac6716\" (UID: \"4047d3e2-47ab-412e-a491-34835eac6716\") " Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.784425 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-logs\") pod \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.784449 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-combined-ca-bundle\") pod \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.784508 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msnkm\" (UniqueName: \"kubernetes.io/projected/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-kube-api-access-msnkm\") pod \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\" (UID: \"29a32f53-8017-4b24-ad95-d20e7ec1aaa0\") " Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.788173 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "29a32f53-8017-4b24-ad95-d20e7ec1aaa0" (UID: "29a32f53-8017-4b24-ad95-d20e7ec1aaa0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.793361 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-logs" (OuterVolumeSpecName: "logs") pod "29a32f53-8017-4b24-ad95-d20e7ec1aaa0" (UID: "29a32f53-8017-4b24-ad95-d20e7ec1aaa0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.793568 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-kube-api-access-msnkm" (OuterVolumeSpecName: "kube-api-access-msnkm") pod "29a32f53-8017-4b24-ad95-d20e7ec1aaa0" (UID: "29a32f53-8017-4b24-ad95-d20e7ec1aaa0"). InnerVolumeSpecName "kube-api-access-msnkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.801731 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-scripts" (OuterVolumeSpecName: "scripts") pod "29a32f53-8017-4b24-ad95-d20e7ec1aaa0" (UID: "29a32f53-8017-4b24-ad95-d20e7ec1aaa0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.803678 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "29a32f53-8017-4b24-ad95-d20e7ec1aaa0" (UID: "29a32f53-8017-4b24-ad95-d20e7ec1aaa0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.803838 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4047d3e2-47ab-412e-a491-34835eac6716-kube-api-access-ppjtf" (OuterVolumeSpecName: "kube-api-access-ppjtf") pod "4047d3e2-47ab-412e-a491-34835eac6716" (UID: "4047d3e2-47ab-412e-a491-34835eac6716"). InnerVolumeSpecName "kube-api-access-ppjtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.852712 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29a32f53-8017-4b24-ad95-d20e7ec1aaa0" (UID: "29a32f53-8017-4b24-ad95-d20e7ec1aaa0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.885828 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.885855 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.885866 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.885875 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msnkm\" (UniqueName: \"kubernetes.io/projected/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-kube-api-access-msnkm\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.885885 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.885892 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.885901 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppjtf\" (UniqueName: \"kubernetes.io/projected/4047d3e2-47ab-412e-a491-34835eac6716-kube-api-access-ppjtf\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.919550 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-config" (OuterVolumeSpecName: "config") pod "4047d3e2-47ab-412e-a491-34835eac6716" (UID: "4047d3e2-47ab-412e-a491-34835eac6716"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.919585 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-config-data" (OuterVolumeSpecName: "config-data") pod "29a32f53-8017-4b24-ad95-d20e7ec1aaa0" (UID: "29a32f53-8017-4b24-ad95-d20e7ec1aaa0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.986744 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:58 crc kubenswrapper[4744]: I1003 16:44:58.986770 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29a32f53-8017-4b24-ad95-d20e7ec1aaa0-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.012876 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4047d3e2-47ab-412e-a491-34835eac6716" (UID: "4047d3e2-47ab-412e-a491-34835eac6716"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.029932 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4047d3e2-47ab-412e-a491-34835eac6716" (UID: "4047d3e2-47ab-412e-a491-34835eac6716"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.030530 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4047d3e2-47ab-412e-a491-34835eac6716" (UID: "4047d3e2-47ab-412e-a491-34835eac6716"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.033012 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4047d3e2-47ab-412e-a491-34835eac6716" (UID: "4047d3e2-47ab-412e-a491-34835eac6716"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.089549 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.089577 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.089590 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.089600 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4047d3e2-47ab-412e-a491-34835eac6716-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.111363 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-wh96r"] Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.218687 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-qsplt"] Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.226288 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-ldg2v"] Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.470909 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 03 16:44:59 crc kubenswrapper[4744]: W1003 16:44:59.476715 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33874615_0e1f_4743_aa84_d262b6eab68c.slice/crio-7aeb304b7edf21d6e6ea160f64e68b6cf057672b0ccae056f2bd67cac60d7544 WatchSource:0}: Error finding container 7aeb304b7edf21d6e6ea160f64e68b6cf057672b0ccae056f2bd67cac60d7544: Status 404 returned error can't find the container with id 7aeb304b7edf21d6e6ea160f64e68b6cf057672b0ccae056f2bd67cac60d7544 Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.508531 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"4fa451ca-3be8-4074-b903-5ea114f47bda","Type":"ContainerStarted","Data":"6292a918a1fbf56b48a41e7b1d948dba59c8a26c8e38e368dfb1779a3c68d8f2"} Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.510125 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1","Type":"ContainerStarted","Data":"55186542a5e8af4fed367e12daa2aa7314b801e5a277e5b28eacd766c40832b2"} Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.511310 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34b80b77-d825-45d7-825c-88a8089fa4ca","Type":"ContainerStarted","Data":"7551ef0e46f97db37f1a03c4bb1785b0590522a2178cb83f9412064278273c19"} Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.514040 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-qsplt" event={"ID":"501a4d70-6aa9-4fd5-8e72-5b854e648219","Type":"ContainerStarted","Data":"9a137d288bf9afcc935660a6c8cc4377050731d231f1dbb19e8a65660eaf297b"} Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.514084 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-qsplt" event={"ID":"501a4d70-6aa9-4fd5-8e72-5b854e648219","Type":"ContainerStarted","Data":"c8f6128e12255e64ddf53d752ba36b72688383411b90876b408da8e1e9a3cc6d"} Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.515906 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"33874615-0e1f-4743-aa84-d262b6eab68c","Type":"ContainerStarted","Data":"7aeb304b7edf21d6e6ea160f64e68b6cf057672b0ccae056f2bd67cac60d7544"} Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.520270 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ldg2v" event={"ID":"15b5b2d9-3e50-4326-a345-8b5f79d684e2","Type":"ContainerStarted","Data":"259b7d4a70bf2684be3f6247908706dd5d345163f48a7840387c842a166e3fb2"} Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.520305 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ldg2v" event={"ID":"15b5b2d9-3e50-4326-a345-8b5f79d684e2","Type":"ContainerStarted","Data":"76b7204d63daa1c2d4caae76403df482b6f5e6dc3edf4e96e268e36c4be6c0df"} Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.531098 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.055021731 podStartE2EDuration="35.531079652s" podCreationTimestamp="2025-10-03 16:44:24 +0000 UTC" firstStartedPulling="2025-10-03 16:44:26.13475103 +0000 UTC m=+1192.414626926" lastFinishedPulling="2025-10-03 16:44:58.610808941 +0000 UTC m=+1224.890684847" observedRunningTime="2025-10-03 16:44:59.529694957 +0000 UTC m=+1225.809570853" watchObservedRunningTime="2025-10-03 16:44:59.531079652 +0000 UTC m=+1225.810955548" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.538973 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"29a32f53-8017-4b24-ad95-d20e7ec1aaa0","Type":"ContainerDied","Data":"974da30332c6d8e4475599f89ca0b8b3f869b05bc210e5b44c88bb2a740679b2"} Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.539028 4744 scope.go:117] "RemoveContainer" containerID="4517f45c77679c3f0d853c515823705bf5abb55c901b955cc81e2fd121d4357a" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.539140 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.560365 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-qsplt" podStartSLOduration=4.560341308 podStartE2EDuration="4.560341308s" podCreationTimestamp="2025-10-03 16:44:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:59.550438805 +0000 UTC m=+1225.830314701" watchObservedRunningTime="2025-10-03 16:44:59.560341308 +0000 UTC m=+1225.840217214" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.566140 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" event={"ID":"4047d3e2-47ab-412e-a491-34835eac6716","Type":"ContainerDied","Data":"18765538cc9146ab7f8bf07ac989eb94b51b91dc70cfc4c2c4d8a6bfb4a8e236"} Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.566237 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-qcrs4" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.577509 4744 generic.go:334] "Generic (PLEG): container finished" podID="cca8b080-5315-4c0e-ab9d-c94c248df68d" containerID="af81ada8db847b37ff88c6b34fd623097cd5f8b6bd7f1cac3c532aa7ca4eb1a3" exitCode=0 Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.577552 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wh96r" event={"ID":"cca8b080-5315-4c0e-ab9d-c94c248df68d","Type":"ContainerDied","Data":"af81ada8db847b37ff88c6b34fd623097cd5f8b6bd7f1cac3c532aa7ca4eb1a3"} Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.577580 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wh96r" event={"ID":"cca8b080-5315-4c0e-ab9d-c94c248df68d","Type":"ContainerStarted","Data":"b78be9b5f4c17c56f3d8af2b0cb50b2c79bcfb2bad418031b9acd23b17ba2f98"} Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.600545 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-ldg2v" podStartSLOduration=4.600527351 podStartE2EDuration="4.600527351s" podCreationTimestamp="2025-10-03 16:44:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:44:59.575448682 +0000 UTC m=+1225.855324588" watchObservedRunningTime="2025-10-03 16:44:59.600527351 +0000 UTC m=+1225.880403237" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.828877 4744 scope.go:117] "RemoveContainer" containerID="075ed500d2422c1e0176fa45154326ae243f99953a80fd5d7e971650fedaad44" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.857234 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.873030 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.874531 4744 scope.go:117] "RemoveContainer" containerID="6ecd7f73349ba82f8168f463061e4ccaea47660f8e0a3feb08119ad164a7be4e" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.887997 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qcrs4"] Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.897568 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qcrs4"] Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.908351 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 03 16:44:59 crc kubenswrapper[4744]: E1003 16:44:59.911043 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a32f53-8017-4b24-ad95-d20e7ec1aaa0" containerName="manila-api" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.911070 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a32f53-8017-4b24-ad95-d20e7ec1aaa0" containerName="manila-api" Oct 03 16:44:59 crc kubenswrapper[4744]: E1003 16:44:59.911088 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a32f53-8017-4b24-ad95-d20e7ec1aaa0" containerName="manila-api-log" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.911094 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a32f53-8017-4b24-ad95-d20e7ec1aaa0" containerName="manila-api-log" Oct 03 16:44:59 crc kubenswrapper[4744]: E1003 16:44:59.911106 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4047d3e2-47ab-412e-a491-34835eac6716" containerName="dnsmasq-dns" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.911112 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4047d3e2-47ab-412e-a491-34835eac6716" containerName="dnsmasq-dns" Oct 03 16:44:59 crc kubenswrapper[4744]: E1003 16:44:59.911123 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4047d3e2-47ab-412e-a491-34835eac6716" containerName="init" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.911128 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4047d3e2-47ab-412e-a491-34835eac6716" containerName="init" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.911295 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a32f53-8017-4b24-ad95-d20e7ec1aaa0" containerName="manila-api" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.911312 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4047d3e2-47ab-412e-a491-34835eac6716" containerName="dnsmasq-dns" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.911330 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a32f53-8017-4b24-ad95-d20e7ec1aaa0" containerName="manila-api-log" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.913097 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.916317 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.916557 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.916716 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.925144 4744 scope.go:117] "RemoveContainer" containerID="fa2048dfd084a40a56d00dda9a6959069f1c98cb6079c278fac7e4c00665658c" Oct 03 16:44:59 crc kubenswrapper[4744]: I1003 16:44:59.935375 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.022458 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc2nv\" (UniqueName: \"kubernetes.io/projected/f392aa4a-226f-4009-acbb-d66b4aaeb95e-kube-api-access-kc2nv\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.022574 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-public-tls-certs\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.022604 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-scripts\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.022638 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f392aa4a-226f-4009-acbb-d66b4aaeb95e-etc-machine-id\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.022737 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.022759 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f392aa4a-226f-4009-acbb-d66b4aaeb95e-logs\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.022783 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-config-data\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.022835 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-config-data-custom\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.022869 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-internal-tls-certs\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.124698 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.124756 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f392aa4a-226f-4009-acbb-d66b4aaeb95e-logs\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.124779 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-config-data\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.124836 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-config-data-custom\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.124875 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-internal-tls-certs\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.124926 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc2nv\" (UniqueName: \"kubernetes.io/projected/f392aa4a-226f-4009-acbb-d66b4aaeb95e-kube-api-access-kc2nv\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.124966 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-public-tls-certs\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.124986 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-scripts\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.125018 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f392aa4a-226f-4009-acbb-d66b4aaeb95e-etc-machine-id\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.125141 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f392aa4a-226f-4009-acbb-d66b4aaeb95e-etc-machine-id\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.134853 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f392aa4a-226f-4009-acbb-d66b4aaeb95e-logs\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.142136 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.142679 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-scripts\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.143789 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr"] Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.152116 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-config-data-custom\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.153922 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.162547 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-config-data\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.164760 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.164810 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.165196 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-public-tls-certs\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.175413 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f392aa4a-226f-4009-acbb-d66b4aaeb95e-internal-tls-certs\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.187433 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc2nv\" (UniqueName: \"kubernetes.io/projected/f392aa4a-226f-4009-acbb-d66b4aaeb95e-kube-api-access-kc2nv\") pod \"manila-api-0\" (UID: \"f392aa4a-226f-4009-acbb-d66b4aaeb95e\") " pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.191798 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr"] Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.241292 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.330505 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-secret-volume\") pod \"collect-profiles-29325165-bxpsr\" (UID: \"451ef308-a0a4-4cf2-9132-7a6c40a9b35c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.330750 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2599b\" (UniqueName: \"kubernetes.io/projected/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-kube-api-access-2599b\") pod \"collect-profiles-29325165-bxpsr\" (UID: \"451ef308-a0a4-4cf2-9132-7a6c40a9b35c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.330806 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-config-volume\") pod \"collect-profiles-29325165-bxpsr\" (UID: \"451ef308-a0a4-4cf2-9132-7a6c40a9b35c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.432573 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-config-volume\") pod \"collect-profiles-29325165-bxpsr\" (UID: \"451ef308-a0a4-4cf2-9132-7a6c40a9b35c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.432764 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-secret-volume\") pod \"collect-profiles-29325165-bxpsr\" (UID: \"451ef308-a0a4-4cf2-9132-7a6c40a9b35c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.432791 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2599b\" (UniqueName: \"kubernetes.io/projected/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-kube-api-access-2599b\") pod \"collect-profiles-29325165-bxpsr\" (UID: \"451ef308-a0a4-4cf2-9132-7a6c40a9b35c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.434074 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-config-volume\") pod \"collect-profiles-29325165-bxpsr\" (UID: \"451ef308-a0a4-4cf2-9132-7a6c40a9b35c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.446862 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-secret-volume\") pod \"collect-profiles-29325165-bxpsr\" (UID: \"451ef308-a0a4-4cf2-9132-7a6c40a9b35c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.454404 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2599b\" (UniqueName: \"kubernetes.io/projected/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-kube-api-access-2599b\") pod \"collect-profiles-29325165-bxpsr\" (UID: \"451ef308-a0a4-4cf2-9132-7a6c40a9b35c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.532571 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.644051 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"4fa451ca-3be8-4074-b903-5ea114f47bda","Type":"ContainerStarted","Data":"53ac3655289c7648b5a9b1b7f77edd67983324dc05b6862f0ec695326ada3bb9"} Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.679475 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34b80b77-d825-45d7-825c-88a8089fa4ca","Type":"ContainerStarted","Data":"d651ca10ea496df1dbb13663510e524697fe7945b63006d97c053f448936070c"} Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.691568 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=4.54793534 podStartE2EDuration="18.691545262s" podCreationTimestamp="2025-10-03 16:44:42 +0000 UTC" firstStartedPulling="2025-10-03 16:44:44.487696971 +0000 UTC m=+1210.767572867" lastFinishedPulling="2025-10-03 16:44:58.631306893 +0000 UTC m=+1224.911182789" observedRunningTime="2025-10-03 16:45:00.677117724 +0000 UTC m=+1226.956993620" watchObservedRunningTime="2025-10-03 16:45:00.691545262 +0000 UTC m=+1226.971421158" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.695786 4744 generic.go:334] "Generic (PLEG): container finished" podID="501a4d70-6aa9-4fd5-8e72-5b854e648219" containerID="9a137d288bf9afcc935660a6c8cc4377050731d231f1dbb19e8a65660eaf297b" exitCode=0 Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.695894 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-qsplt" event={"ID":"501a4d70-6aa9-4fd5-8e72-5b854e648219","Type":"ContainerDied","Data":"9a137d288bf9afcc935660a6c8cc4377050731d231f1dbb19e8a65660eaf297b"} Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.702181 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"33874615-0e1f-4743-aa84-d262b6eab68c","Type":"ContainerStarted","Data":"8ae3f09947e2bcd083bc9683576ab9c2885f273e32221a5f2ed5c436e5ab07bf"} Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.702262 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"33874615-0e1f-4743-aa84-d262b6eab68c","Type":"ContainerStarted","Data":"bba1b6b4bee57c0949e5fd0daaad644d60dfc9adeb84a62be71090a38d374f1a"} Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.724540 4744 generic.go:334] "Generic (PLEG): container finished" podID="15b5b2d9-3e50-4326-a345-8b5f79d684e2" containerID="259b7d4a70bf2684be3f6247908706dd5d345163f48a7840387c842a166e3fb2" exitCode=0 Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.725572 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ldg2v" event={"ID":"15b5b2d9-3e50-4326-a345-8b5f79d684e2","Type":"ContainerDied","Data":"259b7d4a70bf2684be3f6247908706dd5d345163f48a7840387c842a166e3fb2"} Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.817671 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=13.817646083 podStartE2EDuration="13.817646083s" podCreationTimestamp="2025-10-03 16:44:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:45:00.75745686 +0000 UTC m=+1227.037332756" watchObservedRunningTime="2025-10-03 16:45:00.817646083 +0000 UTC m=+1227.097521979" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.879057 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.910076 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29a32f53-8017-4b24-ad95-d20e7ec1aaa0" path="/var/lib/kubelet/pods/29a32f53-8017-4b24-ad95-d20e7ec1aaa0/volumes" Oct 03 16:45:00 crc kubenswrapper[4744]: I1003 16:45:00.910829 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4047d3e2-47ab-412e-a491-34835eac6716" path="/var/lib/kubelet/pods/4047d3e2-47ab-412e-a491-34835eac6716/volumes" Oct 03 16:45:01 crc kubenswrapper[4744]: I1003 16:45:01.093791 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr"] Oct 03 16:45:01 crc kubenswrapper[4744]: I1003 16:45:01.094985 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wh96r" Oct 03 16:45:01 crc kubenswrapper[4744]: I1003 16:45:01.248380 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzl6s\" (UniqueName: \"kubernetes.io/projected/cca8b080-5315-4c0e-ab9d-c94c248df68d-kube-api-access-kzl6s\") pod \"cca8b080-5315-4c0e-ab9d-c94c248df68d\" (UID: \"cca8b080-5315-4c0e-ab9d-c94c248df68d\") " Oct 03 16:45:01 crc kubenswrapper[4744]: I1003 16:45:01.257945 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cca8b080-5315-4c0e-ab9d-c94c248df68d-kube-api-access-kzl6s" (OuterVolumeSpecName: "kube-api-access-kzl6s") pod "cca8b080-5315-4c0e-ab9d-c94c248df68d" (UID: "cca8b080-5315-4c0e-ab9d-c94c248df68d"). InnerVolumeSpecName "kube-api-access-kzl6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:01 crc kubenswrapper[4744]: I1003 16:45:01.351389 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzl6s\" (UniqueName: \"kubernetes.io/projected/cca8b080-5315-4c0e-ab9d-c94c248df68d-kube-api-access-kzl6s\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:01 crc kubenswrapper[4744]: I1003 16:45:01.734567 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f392aa4a-226f-4009-acbb-d66b4aaeb95e","Type":"ContainerStarted","Data":"afafb72f6131a28eae2585462aa15a47db09c6be022943c2a040a9ed90686120"} Oct 03 16:45:01 crc kubenswrapper[4744]: I1003 16:45:01.734618 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f392aa4a-226f-4009-acbb-d66b4aaeb95e","Type":"ContainerStarted","Data":"b933c6e7ff5b69cd325dfd8b24859c1e9189f8a6126160e761ad61b33cc09941"} Oct 03 16:45:01 crc kubenswrapper[4744]: I1003 16:45:01.736170 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wh96r" event={"ID":"cca8b080-5315-4c0e-ab9d-c94c248df68d","Type":"ContainerDied","Data":"b78be9b5f4c17c56f3d8af2b0cb50b2c79bcfb2bad418031b9acd23b17ba2f98"} Oct 03 16:45:01 crc kubenswrapper[4744]: I1003 16:45:01.736200 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b78be9b5f4c17c56f3d8af2b0cb50b2c79bcfb2bad418031b9acd23b17ba2f98" Oct 03 16:45:01 crc kubenswrapper[4744]: I1003 16:45:01.736255 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wh96r" Oct 03 16:45:01 crc kubenswrapper[4744]: I1003 16:45:01.741992 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34b80b77-d825-45d7-825c-88a8089fa4ca","Type":"ContainerStarted","Data":"d92fc025cccb693d4532a96d760686813ed75e6996cf6e3ac0bbf200aaa3e189"} Oct 03 16:45:01 crc kubenswrapper[4744]: I1003 16:45:01.748004 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" event={"ID":"451ef308-a0a4-4cf2-9132-7a6c40a9b35c","Type":"ContainerStarted","Data":"c7fdb9ddea7fb74ac8e2983a1b1fa09376f3576728dbe4eb2e71fbd464568fac"} Oct 03 16:45:01 crc kubenswrapper[4744]: I1003 16:45:01.748072 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" event={"ID":"451ef308-a0a4-4cf2-9132-7a6c40a9b35c","Type":"ContainerStarted","Data":"ae8563599a10521c75a83ecac8f599b4b46fec5ed2bc40e4808918160200fac6"} Oct 03 16:45:01 crc kubenswrapper[4744]: I1003 16:45:01.771005 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" podStartSLOduration=1.770983668 podStartE2EDuration="1.770983668s" podCreationTimestamp="2025-10-03 16:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:45:01.764534564 +0000 UTC m=+1228.044410460" watchObservedRunningTime="2025-10-03 16:45:01.770983668 +0000 UTC m=+1228.050859564" Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.351172 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ldg2v" Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.367688 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qsplt" Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.502261 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjq6l\" (UniqueName: \"kubernetes.io/projected/15b5b2d9-3e50-4326-a345-8b5f79d684e2-kube-api-access-bjq6l\") pod \"15b5b2d9-3e50-4326-a345-8b5f79d684e2\" (UID: \"15b5b2d9-3e50-4326-a345-8b5f79d684e2\") " Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.502691 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsxv9\" (UniqueName: \"kubernetes.io/projected/501a4d70-6aa9-4fd5-8e72-5b854e648219-kube-api-access-gsxv9\") pod \"501a4d70-6aa9-4fd5-8e72-5b854e648219\" (UID: \"501a4d70-6aa9-4fd5-8e72-5b854e648219\") " Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.507689 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/501a4d70-6aa9-4fd5-8e72-5b854e648219-kube-api-access-gsxv9" (OuterVolumeSpecName: "kube-api-access-gsxv9") pod "501a4d70-6aa9-4fd5-8e72-5b854e648219" (UID: "501a4d70-6aa9-4fd5-8e72-5b854e648219"). InnerVolumeSpecName "kube-api-access-gsxv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.509625 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15b5b2d9-3e50-4326-a345-8b5f79d684e2-kube-api-access-bjq6l" (OuterVolumeSpecName: "kube-api-access-bjq6l") pod "15b5b2d9-3e50-4326-a345-8b5f79d684e2" (UID: "15b5b2d9-3e50-4326-a345-8b5f79d684e2"). InnerVolumeSpecName "kube-api-access-bjq6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.610441 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjq6l\" (UniqueName: \"kubernetes.io/projected/15b5b2d9-3e50-4326-a345-8b5f79d684e2-kube-api-access-bjq6l\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.610535 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsxv9\" (UniqueName: \"kubernetes.io/projected/501a4d70-6aa9-4fd5-8e72-5b854e648219-kube-api-access-gsxv9\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.717604 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.757418 4744 generic.go:334] "Generic (PLEG): container finished" podID="451ef308-a0a4-4cf2-9132-7a6c40a9b35c" containerID="c7fdb9ddea7fb74ac8e2983a1b1fa09376f3576728dbe4eb2e71fbd464568fac" exitCode=0 Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.757539 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" event={"ID":"451ef308-a0a4-4cf2-9132-7a6c40a9b35c","Type":"ContainerDied","Data":"c7fdb9ddea7fb74ac8e2983a1b1fa09376f3576728dbe4eb2e71fbd464568fac"} Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.759385 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ldg2v" event={"ID":"15b5b2d9-3e50-4326-a345-8b5f79d684e2","Type":"ContainerDied","Data":"76b7204d63daa1c2d4caae76403df482b6f5e6dc3edf4e96e268e36c4be6c0df"} Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.759430 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ldg2v" Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.759461 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76b7204d63daa1c2d4caae76403df482b6f5e6dc3edf4e96e268e36c4be6c0df" Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.761650 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f392aa4a-226f-4009-acbb-d66b4aaeb95e","Type":"ContainerStarted","Data":"9c3ec3457be7b83c2e789e99782c284e9e1712872459be41f78d375c202b1348"} Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.761743 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.763368 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-qsplt" event={"ID":"501a4d70-6aa9-4fd5-8e72-5b854e648219","Type":"ContainerDied","Data":"c8f6128e12255e64ddf53d752ba36b72688383411b90876b408da8e1e9a3cc6d"} Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.763395 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-qsplt" Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.763409 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8f6128e12255e64ddf53d752ba36b72688383411b90876b408da8e1e9a3cc6d" Oct 03 16:45:02 crc kubenswrapper[4744]: I1003 16:45:02.804809 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.804791212 podStartE2EDuration="3.804791212s" podCreationTimestamp="2025-10-03 16:44:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:45:02.795063414 +0000 UTC m=+1229.074939310" watchObservedRunningTime="2025-10-03 16:45:02.804791212 +0000 UTC m=+1229.084667108" Oct 03 16:45:03 crc kubenswrapper[4744]: I1003 16:45:03.510685 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 03 16:45:03 crc kubenswrapper[4744]: I1003 16:45:03.775380 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34b80b77-d825-45d7-825c-88a8089fa4ca","Type":"ContainerStarted","Data":"b64218164704614d9e2138f858964db995358150e6f3ebca1ec228750f0bdb62"} Oct 03 16:45:03 crc kubenswrapper[4744]: I1003 16:45:03.775583 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="ceilometer-central-agent" containerID="cri-o://7551ef0e46f97db37f1a03c4bb1785b0590522a2178cb83f9412064278273c19" gracePeriod=30 Oct 03 16:45:03 crc kubenswrapper[4744]: I1003 16:45:03.775606 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="ceilometer-notification-agent" containerID="cri-o://d651ca10ea496df1dbb13663510e524697fe7945b63006d97c053f448936070c" gracePeriod=30 Oct 03 16:45:03 crc kubenswrapper[4744]: I1003 16:45:03.775598 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="sg-core" containerID="cri-o://d92fc025cccb693d4532a96d760686813ed75e6996cf6e3ac0bbf200aaa3e189" gracePeriod=30 Oct 03 16:45:03 crc kubenswrapper[4744]: I1003 16:45:03.775732 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="proxy-httpd" containerID="cri-o://b64218164704614d9e2138f858964db995358150e6f3ebca1ec228750f0bdb62" gracePeriod=30 Oct 03 16:45:03 crc kubenswrapper[4744]: I1003 16:45:03.775947 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 16:45:03 crc kubenswrapper[4744]: I1003 16:45:03.801807 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=9.955685924 podStartE2EDuration="17.801793098s" podCreationTimestamp="2025-10-03 16:44:46 +0000 UTC" firstStartedPulling="2025-10-03 16:44:55.220040192 +0000 UTC m=+1221.499916088" lastFinishedPulling="2025-10-03 16:45:03.066147366 +0000 UTC m=+1229.346023262" observedRunningTime="2025-10-03 16:45:03.79992657 +0000 UTC m=+1230.079802466" watchObservedRunningTime="2025-10-03 16:45:03.801793098 +0000 UTC m=+1230.081668994" Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.263724 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.345669 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2599b\" (UniqueName: \"kubernetes.io/projected/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-kube-api-access-2599b\") pod \"451ef308-a0a4-4cf2-9132-7a6c40a9b35c\" (UID: \"451ef308-a0a4-4cf2-9132-7a6c40a9b35c\") " Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.345748 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-config-volume\") pod \"451ef308-a0a4-4cf2-9132-7a6c40a9b35c\" (UID: \"451ef308-a0a4-4cf2-9132-7a6c40a9b35c\") " Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.345974 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-secret-volume\") pod \"451ef308-a0a4-4cf2-9132-7a6c40a9b35c\" (UID: \"451ef308-a0a4-4cf2-9132-7a6c40a9b35c\") " Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.347458 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-config-volume" (OuterVolumeSpecName: "config-volume") pod "451ef308-a0a4-4cf2-9132-7a6c40a9b35c" (UID: "451ef308-a0a4-4cf2-9132-7a6c40a9b35c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.351942 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-kube-api-access-2599b" (OuterVolumeSpecName: "kube-api-access-2599b") pod "451ef308-a0a4-4cf2-9132-7a6c40a9b35c" (UID: "451ef308-a0a4-4cf2-9132-7a6c40a9b35c"). InnerVolumeSpecName "kube-api-access-2599b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.366574 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "451ef308-a0a4-4cf2-9132-7a6c40a9b35c" (UID: "451ef308-a0a4-4cf2-9132-7a6c40a9b35c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.449383 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.449420 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2599b\" (UniqueName: \"kubernetes.io/projected/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-kube-api-access-2599b\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.449436 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/451ef308-a0a4-4cf2-9132-7a6c40a9b35c-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.787941 4744 generic.go:334] "Generic (PLEG): container finished" podID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerID="b64218164704614d9e2138f858964db995358150e6f3ebca1ec228750f0bdb62" exitCode=0 Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.788301 4744 generic.go:334] "Generic (PLEG): container finished" podID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerID="d92fc025cccb693d4532a96d760686813ed75e6996cf6e3ac0bbf200aaa3e189" exitCode=2 Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.788317 4744 generic.go:334] "Generic (PLEG): container finished" podID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerID="d651ca10ea496df1dbb13663510e524697fe7945b63006d97c053f448936070c" exitCode=0 Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.788368 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34b80b77-d825-45d7-825c-88a8089fa4ca","Type":"ContainerDied","Data":"b64218164704614d9e2138f858964db995358150e6f3ebca1ec228750f0bdb62"} Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.788400 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34b80b77-d825-45d7-825c-88a8089fa4ca","Type":"ContainerDied","Data":"d92fc025cccb693d4532a96d760686813ed75e6996cf6e3ac0bbf200aaa3e189"} Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.788418 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34b80b77-d825-45d7-825c-88a8089fa4ca","Type":"ContainerDied","Data":"d651ca10ea496df1dbb13663510e524697fe7945b63006d97c053f448936070c"} Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.791135 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" event={"ID":"451ef308-a0a4-4cf2-9132-7a6c40a9b35c","Type":"ContainerDied","Data":"ae8563599a10521c75a83ecac8f599b4b46fec5ed2bc40e4808918160200fac6"} Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.791177 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr" Oct 03 16:45:04 crc kubenswrapper[4744]: I1003 16:45:04.791191 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae8563599a10521c75a83ecac8f599b4b46fec5ed2bc40e4808918160200fac6" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.121728 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-b978-account-create-4wzh7"] Oct 03 16:45:05 crc kubenswrapper[4744]: E1003 16:45:05.122179 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15b5b2d9-3e50-4326-a345-8b5f79d684e2" containerName="mariadb-database-create" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.122195 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="15b5b2d9-3e50-4326-a345-8b5f79d684e2" containerName="mariadb-database-create" Oct 03 16:45:05 crc kubenswrapper[4744]: E1003 16:45:05.122208 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501a4d70-6aa9-4fd5-8e72-5b854e648219" containerName="mariadb-database-create" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.122214 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="501a4d70-6aa9-4fd5-8e72-5b854e648219" containerName="mariadb-database-create" Oct 03 16:45:05 crc kubenswrapper[4744]: E1003 16:45:05.122239 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="451ef308-a0a4-4cf2-9132-7a6c40a9b35c" containerName="collect-profiles" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.122245 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="451ef308-a0a4-4cf2-9132-7a6c40a9b35c" containerName="collect-profiles" Oct 03 16:45:05 crc kubenswrapper[4744]: E1003 16:45:05.122258 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cca8b080-5315-4c0e-ab9d-c94c248df68d" containerName="mariadb-database-create" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.122264 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cca8b080-5315-4c0e-ab9d-c94c248df68d" containerName="mariadb-database-create" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.122430 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="501a4d70-6aa9-4fd5-8e72-5b854e648219" containerName="mariadb-database-create" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.122448 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cca8b080-5315-4c0e-ab9d-c94c248df68d" containerName="mariadb-database-create" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.122459 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="451ef308-a0a4-4cf2-9132-7a6c40a9b35c" containerName="collect-profiles" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.122471 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="15b5b2d9-3e50-4326-a345-8b5f79d684e2" containerName="mariadb-database-create" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.123075 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b978-account-create-4wzh7" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.125125 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.131667 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b978-account-create-4wzh7"] Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.165592 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvxqw\" (UniqueName: \"kubernetes.io/projected/e9826741-7031-4894-8ff1-ad1a150f367a-kube-api-access-xvxqw\") pod \"nova-api-b978-account-create-4wzh7\" (UID: \"e9826741-7031-4894-8ff1-ad1a150f367a\") " pod="openstack/nova-api-b978-account-create-4wzh7" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.168849 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.217551 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.267714 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvxqw\" (UniqueName: \"kubernetes.io/projected/e9826741-7031-4894-8ff1-ad1a150f367a-kube-api-access-xvxqw\") pod \"nova-api-b978-account-create-4wzh7\" (UID: \"e9826741-7031-4894-8ff1-ad1a150f367a\") " pod="openstack/nova-api-b978-account-create-4wzh7" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.297124 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvxqw\" (UniqueName: \"kubernetes.io/projected/e9826741-7031-4894-8ff1-ad1a150f367a-kube-api-access-xvxqw\") pod \"nova-api-b978-account-create-4wzh7\" (UID: \"e9826741-7031-4894-8ff1-ad1a150f367a\") " pod="openstack/nova-api-b978-account-create-4wzh7" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.452987 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b978-account-create-4wzh7" Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.814303 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="5171a06a-cac6-4fa5-abfc-8423d21003ce" containerName="manila-scheduler" containerID="cri-o://e7e73d5de0975ca0698a8f0cd60545f25b1cfea11b338e43e683a3199a0a9036" gracePeriod=30 Oct 03 16:45:05 crc kubenswrapper[4744]: I1003 16:45:05.815437 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="5171a06a-cac6-4fa5-abfc-8423d21003ce" containerName="probe" containerID="cri-o://39ccc5d10f57ae51069eed00f8c31eb82d8557f26f7e79032a711276b096ef2a" gracePeriod=30 Oct 03 16:45:06 crc kubenswrapper[4744]: I1003 16:45:06.033163 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b978-account-create-4wzh7"] Oct 03 16:45:06 crc kubenswrapper[4744]: I1003 16:45:06.824262 4744 generic.go:334] "Generic (PLEG): container finished" podID="e9826741-7031-4894-8ff1-ad1a150f367a" containerID="0ea1cd99b96b79967a4515714aea8d00a35506053339e368c6622b035958e0d4" exitCode=0 Oct 03 16:45:06 crc kubenswrapper[4744]: I1003 16:45:06.824341 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b978-account-create-4wzh7" event={"ID":"e9826741-7031-4894-8ff1-ad1a150f367a","Type":"ContainerDied","Data":"0ea1cd99b96b79967a4515714aea8d00a35506053339e368c6622b035958e0d4"} Oct 03 16:45:06 crc kubenswrapper[4744]: I1003 16:45:06.824373 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b978-account-create-4wzh7" event={"ID":"e9826741-7031-4894-8ff1-ad1a150f367a","Type":"ContainerStarted","Data":"382b573975dd9ed3240f869cf8e2501916ca568ef58869f24d2b760160f657d0"} Oct 03 16:45:06 crc kubenswrapper[4744]: I1003 16:45:06.827277 4744 generic.go:334] "Generic (PLEG): container finished" podID="5171a06a-cac6-4fa5-abfc-8423d21003ce" containerID="39ccc5d10f57ae51069eed00f8c31eb82d8557f26f7e79032a711276b096ef2a" exitCode=0 Oct 03 16:45:06 crc kubenswrapper[4744]: I1003 16:45:06.827325 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5171a06a-cac6-4fa5-abfc-8423d21003ce","Type":"ContainerDied","Data":"39ccc5d10f57ae51069eed00f8c31eb82d8557f26f7e79032a711276b096ef2a"} Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.250731 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.251234 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="cb8803f7-c116-4175-b653-4b451e1da26d" containerName="glance-log" containerID="cri-o://3c3fbe34545f10996e3b7700b35eb0c4743eb42323a910487b6f2b9afae26aaf" gracePeriod=30 Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.251691 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="cb8803f7-c116-4175-b653-4b451e1da26d" containerName="glance-httpd" containerID="cri-o://44bc9bc5472efb50fcdd2da3dcd56942d91f2c82a55c8013f083377f384f0b06" gracePeriod=30 Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.472662 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.543583 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m72pc\" (UniqueName: \"kubernetes.io/projected/34b80b77-d825-45d7-825c-88a8089fa4ca-kube-api-access-m72pc\") pod \"34b80b77-d825-45d7-825c-88a8089fa4ca\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.543701 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34b80b77-d825-45d7-825c-88a8089fa4ca-run-httpd\") pod \"34b80b77-d825-45d7-825c-88a8089fa4ca\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.543809 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-config-data\") pod \"34b80b77-d825-45d7-825c-88a8089fa4ca\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.543936 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34b80b77-d825-45d7-825c-88a8089fa4ca-log-httpd\") pod \"34b80b77-d825-45d7-825c-88a8089fa4ca\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.543991 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-combined-ca-bundle\") pod \"34b80b77-d825-45d7-825c-88a8089fa4ca\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.544034 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-scripts\") pod \"34b80b77-d825-45d7-825c-88a8089fa4ca\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.544069 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-sg-core-conf-yaml\") pod \"34b80b77-d825-45d7-825c-88a8089fa4ca\" (UID: \"34b80b77-d825-45d7-825c-88a8089fa4ca\") " Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.544415 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34b80b77-d825-45d7-825c-88a8089fa4ca-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "34b80b77-d825-45d7-825c-88a8089fa4ca" (UID: "34b80b77-d825-45d7-825c-88a8089fa4ca"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.545863 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34b80b77-d825-45d7-825c-88a8089fa4ca-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "34b80b77-d825-45d7-825c-88a8089fa4ca" (UID: "34b80b77-d825-45d7-825c-88a8089fa4ca"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.550600 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34b80b77-d825-45d7-825c-88a8089fa4ca-kube-api-access-m72pc" (OuterVolumeSpecName: "kube-api-access-m72pc") pod "34b80b77-d825-45d7-825c-88a8089fa4ca" (UID: "34b80b77-d825-45d7-825c-88a8089fa4ca"). InnerVolumeSpecName "kube-api-access-m72pc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.550895 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-scripts" (OuterVolumeSpecName: "scripts") pod "34b80b77-d825-45d7-825c-88a8089fa4ca" (UID: "34b80b77-d825-45d7-825c-88a8089fa4ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.601572 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "34b80b77-d825-45d7-825c-88a8089fa4ca" (UID: "34b80b77-d825-45d7-825c-88a8089fa4ca"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.642086 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34b80b77-d825-45d7-825c-88a8089fa4ca" (UID: "34b80b77-d825-45d7-825c-88a8089fa4ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.646181 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34b80b77-d825-45d7-825c-88a8089fa4ca-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.646341 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.646412 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.646483 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.646570 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m72pc\" (UniqueName: \"kubernetes.io/projected/34b80b77-d825-45d7-825c-88a8089fa4ca-kube-api-access-m72pc\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.646626 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34b80b77-d825-45d7-825c-88a8089fa4ca-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.678700 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-config-data" (OuterVolumeSpecName: "config-data") pod "34b80b77-d825-45d7-825c-88a8089fa4ca" (UID: "34b80b77-d825-45d7-825c-88a8089fa4ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.752401 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b80b77-d825-45d7-825c-88a8089fa4ca-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.844801 4744 generic.go:334] "Generic (PLEG): container finished" podID="5171a06a-cac6-4fa5-abfc-8423d21003ce" containerID="e7e73d5de0975ca0698a8f0cd60545f25b1cfea11b338e43e683a3199a0a9036" exitCode=0 Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.844871 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5171a06a-cac6-4fa5-abfc-8423d21003ce","Type":"ContainerDied","Data":"e7e73d5de0975ca0698a8f0cd60545f25b1cfea11b338e43e683a3199a0a9036"} Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.848054 4744 generic.go:334] "Generic (PLEG): container finished" podID="cb8803f7-c116-4175-b653-4b451e1da26d" containerID="3c3fbe34545f10996e3b7700b35eb0c4743eb42323a910487b6f2b9afae26aaf" exitCode=143 Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.848117 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb8803f7-c116-4175-b653-4b451e1da26d","Type":"ContainerDied","Data":"3c3fbe34545f10996e3b7700b35eb0c4743eb42323a910487b6f2b9afae26aaf"} Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.853961 4744 generic.go:334] "Generic (PLEG): container finished" podID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerID="7551ef0e46f97db37f1a03c4bb1785b0590522a2178cb83f9412064278273c19" exitCode=0 Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.854054 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.854068 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34b80b77-d825-45d7-825c-88a8089fa4ca","Type":"ContainerDied","Data":"7551ef0e46f97db37f1a03c4bb1785b0590522a2178cb83f9412064278273c19"} Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.854145 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"34b80b77-d825-45d7-825c-88a8089fa4ca","Type":"ContainerDied","Data":"5d2258829ece1902f0ebc0685a6d055a1a6a3b65a93fe1af30b94ee9d4547789"} Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.854172 4744 scope.go:117] "RemoveContainer" containerID="b64218164704614d9e2138f858964db995358150e6f3ebca1ec228750f0bdb62" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.882364 4744 scope.go:117] "RemoveContainer" containerID="d92fc025cccb693d4532a96d760686813ed75e6996cf6e3ac0bbf200aaa3e189" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.917324 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.964486 4744 scope.go:117] "RemoveContainer" containerID="d651ca10ea496df1dbb13663510e524697fe7945b63006d97c053f448936070c" Oct 03 16:45:07 crc kubenswrapper[4744]: I1003 16:45:07.966052 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.014358 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:08 crc kubenswrapper[4744]: E1003 16:45:08.014825 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="ceilometer-notification-agent" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.014846 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="ceilometer-notification-agent" Oct 03 16:45:08 crc kubenswrapper[4744]: E1003 16:45:08.014953 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="proxy-httpd" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.014988 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="proxy-httpd" Oct 03 16:45:08 crc kubenswrapper[4744]: E1003 16:45:08.015011 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="sg-core" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.015017 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="sg-core" Oct 03 16:45:08 crc kubenswrapper[4744]: E1003 16:45:08.015036 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="ceilometer-central-agent" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.015042 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="ceilometer-central-agent" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.015374 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="ceilometer-notification-agent" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.015386 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="proxy-httpd" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.015404 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="ceilometer-central-agent" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.015416 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" containerName="sg-core" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.017552 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.020212 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.020398 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.031932 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.037533 4744 scope.go:117] "RemoveContainer" containerID="7551ef0e46f97db37f1a03c4bb1785b0590522a2178cb83f9412064278273c19" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.038606 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.058542 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:08 crc kubenswrapper[4744]: E1003 16:45:08.059366 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-k7wbh log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[combined-ca-bundle config-data kube-api-access-k7wbh log-httpd run-httpd scripts sg-core-conf-yaml]: context canceled" pod="openstack/ceilometer-0" podUID="393a8d5d-f3b9-4846-ae16-afe78f29237e" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.078972 4744 scope.go:117] "RemoveContainer" containerID="b64218164704614d9e2138f858964db995358150e6f3ebca1ec228750f0bdb62" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.079218 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 03 16:45:08 crc kubenswrapper[4744]: E1003 16:45:08.082530 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b64218164704614d9e2138f858964db995358150e6f3ebca1ec228750f0bdb62\": container with ID starting with b64218164704614d9e2138f858964db995358150e6f3ebca1ec228750f0bdb62 not found: ID does not exist" containerID="b64218164704614d9e2138f858964db995358150e6f3ebca1ec228750f0bdb62" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.082833 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b64218164704614d9e2138f858964db995358150e6f3ebca1ec228750f0bdb62"} err="failed to get container status \"b64218164704614d9e2138f858964db995358150e6f3ebca1ec228750f0bdb62\": rpc error: code = NotFound desc = could not find container \"b64218164704614d9e2138f858964db995358150e6f3ebca1ec228750f0bdb62\": container with ID starting with b64218164704614d9e2138f858964db995358150e6f3ebca1ec228750f0bdb62 not found: ID does not exist" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.082865 4744 scope.go:117] "RemoveContainer" containerID="d92fc025cccb693d4532a96d760686813ed75e6996cf6e3ac0bbf200aaa3e189" Oct 03 16:45:08 crc kubenswrapper[4744]: E1003 16:45:08.083164 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d92fc025cccb693d4532a96d760686813ed75e6996cf6e3ac0bbf200aaa3e189\": container with ID starting with d92fc025cccb693d4532a96d760686813ed75e6996cf6e3ac0bbf200aaa3e189 not found: ID does not exist" containerID="d92fc025cccb693d4532a96d760686813ed75e6996cf6e3ac0bbf200aaa3e189" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.083244 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d92fc025cccb693d4532a96d760686813ed75e6996cf6e3ac0bbf200aaa3e189"} err="failed to get container status \"d92fc025cccb693d4532a96d760686813ed75e6996cf6e3ac0bbf200aaa3e189\": rpc error: code = NotFound desc = could not find container \"d92fc025cccb693d4532a96d760686813ed75e6996cf6e3ac0bbf200aaa3e189\": container with ID starting with d92fc025cccb693d4532a96d760686813ed75e6996cf6e3ac0bbf200aaa3e189 not found: ID does not exist" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.083307 4744 scope.go:117] "RemoveContainer" containerID="d651ca10ea496df1dbb13663510e524697fe7945b63006d97c053f448936070c" Oct 03 16:45:08 crc kubenswrapper[4744]: E1003 16:45:08.083545 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d651ca10ea496df1dbb13663510e524697fe7945b63006d97c053f448936070c\": container with ID starting with d651ca10ea496df1dbb13663510e524697fe7945b63006d97c053f448936070c not found: ID does not exist" containerID="d651ca10ea496df1dbb13663510e524697fe7945b63006d97c053f448936070c" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.083753 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d651ca10ea496df1dbb13663510e524697fe7945b63006d97c053f448936070c"} err="failed to get container status \"d651ca10ea496df1dbb13663510e524697fe7945b63006d97c053f448936070c\": rpc error: code = NotFound desc = could not find container \"d651ca10ea496df1dbb13663510e524697fe7945b63006d97c053f448936070c\": container with ID starting with d651ca10ea496df1dbb13663510e524697fe7945b63006d97c053f448936070c not found: ID does not exist" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.083826 4744 scope.go:117] "RemoveContainer" containerID="7551ef0e46f97db37f1a03c4bb1785b0590522a2178cb83f9412064278273c19" Oct 03 16:45:08 crc kubenswrapper[4744]: E1003 16:45:08.084596 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7551ef0e46f97db37f1a03c4bb1785b0590522a2178cb83f9412064278273c19\": container with ID starting with 7551ef0e46f97db37f1a03c4bb1785b0590522a2178cb83f9412064278273c19 not found: ID does not exist" containerID="7551ef0e46f97db37f1a03c4bb1785b0590522a2178cb83f9412064278273c19" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.084701 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7551ef0e46f97db37f1a03c4bb1785b0590522a2178cb83f9412064278273c19"} err="failed to get container status \"7551ef0e46f97db37f1a03c4bb1785b0590522a2178cb83f9412064278273c19\": rpc error: code = NotFound desc = could not find container \"7551ef0e46f97db37f1a03c4bb1785b0590522a2178cb83f9412064278273c19\": container with ID starting with 7551ef0e46f97db37f1a03c4bb1785b0590522a2178cb83f9412064278273c19 not found: ID does not exist" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.181869 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjcvh\" (UniqueName: \"kubernetes.io/projected/5171a06a-cac6-4fa5-abfc-8423d21003ce-kube-api-access-mjcvh\") pod \"5171a06a-cac6-4fa5-abfc-8423d21003ce\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.181924 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-config-data\") pod \"5171a06a-cac6-4fa5-abfc-8423d21003ce\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.181954 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-scripts\") pod \"5171a06a-cac6-4fa5-abfc-8423d21003ce\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.182015 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-combined-ca-bundle\") pod \"5171a06a-cac6-4fa5-abfc-8423d21003ce\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.182206 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5171a06a-cac6-4fa5-abfc-8423d21003ce-etc-machine-id\") pod \"5171a06a-cac6-4fa5-abfc-8423d21003ce\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.182283 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-config-data-custom\") pod \"5171a06a-cac6-4fa5-abfc-8423d21003ce\" (UID: \"5171a06a-cac6-4fa5-abfc-8423d21003ce\") " Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.182605 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-scripts\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.182637 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.182685 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7wbh\" (UniqueName: \"kubernetes.io/projected/393a8d5d-f3b9-4846-ae16-afe78f29237e-kube-api-access-k7wbh\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.182765 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.182892 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-config-data\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.183006 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/393a8d5d-f3b9-4846-ae16-afe78f29237e-run-httpd\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.183029 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/393a8d5d-f3b9-4846-ae16-afe78f29237e-log-httpd\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.182670 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5171a06a-cac6-4fa5-abfc-8423d21003ce-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5171a06a-cac6-4fa5-abfc-8423d21003ce" (UID: "5171a06a-cac6-4fa5-abfc-8423d21003ce"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.187702 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5171a06a-cac6-4fa5-abfc-8423d21003ce-kube-api-access-mjcvh" (OuterVolumeSpecName: "kube-api-access-mjcvh") pod "5171a06a-cac6-4fa5-abfc-8423d21003ce" (UID: "5171a06a-cac6-4fa5-abfc-8423d21003ce"). InnerVolumeSpecName "kube-api-access-mjcvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.188016 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-scripts" (OuterVolumeSpecName: "scripts") pod "5171a06a-cac6-4fa5-abfc-8423d21003ce" (UID: "5171a06a-cac6-4fa5-abfc-8423d21003ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.191281 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5171a06a-cac6-4fa5-abfc-8423d21003ce" (UID: "5171a06a-cac6-4fa5-abfc-8423d21003ce"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.253100 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5171a06a-cac6-4fa5-abfc-8423d21003ce" (UID: "5171a06a-cac6-4fa5-abfc-8423d21003ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.270415 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b978-account-create-4wzh7" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.291014 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/393a8d5d-f3b9-4846-ae16-afe78f29237e-run-httpd\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.291052 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/393a8d5d-f3b9-4846-ae16-afe78f29237e-log-httpd\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.291128 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-scripts\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.291147 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.291172 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7wbh\" (UniqueName: \"kubernetes.io/projected/393a8d5d-f3b9-4846-ae16-afe78f29237e-kube-api-access-k7wbh\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.291195 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.291249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-config-data\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.291325 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5171a06a-cac6-4fa5-abfc-8423d21003ce-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.291336 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.291347 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjcvh\" (UniqueName: \"kubernetes.io/projected/5171a06a-cac6-4fa5-abfc-8423d21003ce-kube-api-access-mjcvh\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.291356 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.291366 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.295045 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-config-data\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.300520 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/393a8d5d-f3b9-4846-ae16-afe78f29237e-run-httpd\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.301138 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/393a8d5d-f3b9-4846-ae16-afe78f29237e-log-httpd\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.303547 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-scripts\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.304122 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.311069 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-config-data" (OuterVolumeSpecName: "config-data") pod "5171a06a-cac6-4fa5-abfc-8423d21003ce" (UID: "5171a06a-cac6-4fa5-abfc-8423d21003ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.311586 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.317684 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7wbh\" (UniqueName: \"kubernetes.io/projected/393a8d5d-f3b9-4846-ae16-afe78f29237e-kube-api-access-k7wbh\") pod \"ceilometer-0\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.392888 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvxqw\" (UniqueName: \"kubernetes.io/projected/e9826741-7031-4894-8ff1-ad1a150f367a-kube-api-access-xvxqw\") pod \"e9826741-7031-4894-8ff1-ad1a150f367a\" (UID: \"e9826741-7031-4894-8ff1-ad1a150f367a\") " Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.393405 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5171a06a-cac6-4fa5-abfc-8423d21003ce-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.401196 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9826741-7031-4894-8ff1-ad1a150f367a-kube-api-access-xvxqw" (OuterVolumeSpecName: "kube-api-access-xvxqw") pod "e9826741-7031-4894-8ff1-ad1a150f367a" (UID: "e9826741-7031-4894-8ff1-ad1a150f367a"). InnerVolumeSpecName "kube-api-access-xvxqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.495513 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvxqw\" (UniqueName: \"kubernetes.io/projected/e9826741-7031-4894-8ff1-ad1a150f367a-kube-api-access-xvxqw\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.864963 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b978-account-create-4wzh7" event={"ID":"e9826741-7031-4894-8ff1-ad1a150f367a","Type":"ContainerDied","Data":"382b573975dd9ed3240f869cf8e2501916ca568ef58869f24d2b760160f657d0"} Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.865003 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="382b573975dd9ed3240f869cf8e2501916ca568ef58869f24d2b760160f657d0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.865050 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b978-account-create-4wzh7" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.872234 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5171a06a-cac6-4fa5-abfc-8423d21003ce","Type":"ContainerDied","Data":"a62d6b537ef68100ec667232e02233821286e60b44bc353ee01fd24d8a68f4bc"} Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.872308 4744 scope.go:117] "RemoveContainer" containerID="39ccc5d10f57ae51069eed00f8c31eb82d8557f26f7e79032a711276b096ef2a" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.872445 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.875478 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.893774 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.903031 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34b80b77-d825-45d7-825c-88a8089fa4ca" path="/var/lib/kubelet/pods/34b80b77-d825-45d7-825c-88a8089fa4ca/volumes" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.930681 4744 scope.go:117] "RemoveContainer" containerID="e7e73d5de0975ca0698a8f0cd60545f25b1cfea11b338e43e683a3199a0a9036" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.945043 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.964659 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.987128 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 03 16:45:08 crc kubenswrapper[4744]: E1003 16:45:08.987830 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5171a06a-cac6-4fa5-abfc-8423d21003ce" containerName="probe" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.987864 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5171a06a-cac6-4fa5-abfc-8423d21003ce" containerName="probe" Oct 03 16:45:08 crc kubenswrapper[4744]: E1003 16:45:08.987901 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5171a06a-cac6-4fa5-abfc-8423d21003ce" containerName="manila-scheduler" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.987908 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5171a06a-cac6-4fa5-abfc-8423d21003ce" containerName="manila-scheduler" Oct 03 16:45:08 crc kubenswrapper[4744]: E1003 16:45:08.987924 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9826741-7031-4894-8ff1-ad1a150f367a" containerName="mariadb-account-create" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.987931 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9826741-7031-4894-8ff1-ad1a150f367a" containerName="mariadb-account-create" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.988101 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5171a06a-cac6-4fa5-abfc-8423d21003ce" containerName="probe" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.988118 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5171a06a-cac6-4fa5-abfc-8423d21003ce" containerName="manila-scheduler" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.988127 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9826741-7031-4894-8ff1-ad1a150f367a" containerName="mariadb-account-create" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.989212 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 03 16:45:08 crc kubenswrapper[4744]: I1003 16:45:08.991229 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.005921 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/393a8d5d-f3b9-4846-ae16-afe78f29237e-run-httpd\") pod \"393a8d5d-f3b9-4846-ae16-afe78f29237e\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.005970 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7wbh\" (UniqueName: \"kubernetes.io/projected/393a8d5d-f3b9-4846-ae16-afe78f29237e-kube-api-access-k7wbh\") pod \"393a8d5d-f3b9-4846-ae16-afe78f29237e\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.006326 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/393a8d5d-f3b9-4846-ae16-afe78f29237e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "393a8d5d-f3b9-4846-ae16-afe78f29237e" (UID: "393a8d5d-f3b9-4846-ae16-afe78f29237e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.007125 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-config-data\") pod \"393a8d5d-f3b9-4846-ae16-afe78f29237e\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.011062 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/393a8d5d-f3b9-4846-ae16-afe78f29237e-kube-api-access-k7wbh" (OuterVolumeSpecName: "kube-api-access-k7wbh") pod "393a8d5d-f3b9-4846-ae16-afe78f29237e" (UID: "393a8d5d-f3b9-4846-ae16-afe78f29237e"). InnerVolumeSpecName "kube-api-access-k7wbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.011758 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-combined-ca-bundle\") pod \"393a8d5d-f3b9-4846-ae16-afe78f29237e\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.011921 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/393a8d5d-f3b9-4846-ae16-afe78f29237e-log-httpd\") pod \"393a8d5d-f3b9-4846-ae16-afe78f29237e\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.011977 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-scripts\") pod \"393a8d5d-f3b9-4846-ae16-afe78f29237e\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.012035 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-sg-core-conf-yaml\") pod \"393a8d5d-f3b9-4846-ae16-afe78f29237e\" (UID: \"393a8d5d-f3b9-4846-ae16-afe78f29237e\") " Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.012269 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/393a8d5d-f3b9-4846-ae16-afe78f29237e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "393a8d5d-f3b9-4846-ae16-afe78f29237e" (UID: "393a8d5d-f3b9-4846-ae16-afe78f29237e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.014663 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.014804 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/393a8d5d-f3b9-4846-ae16-afe78f29237e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.014827 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/393a8d5d-f3b9-4846-ae16-afe78f29237e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.014838 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7wbh\" (UniqueName: \"kubernetes.io/projected/393a8d5d-f3b9-4846-ae16-afe78f29237e-kube-api-access-k7wbh\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.015038 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-config-data" (OuterVolumeSpecName: "config-data") pod "393a8d5d-f3b9-4846-ae16-afe78f29237e" (UID: "393a8d5d-f3b9-4846-ae16-afe78f29237e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.015060 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "393a8d5d-f3b9-4846-ae16-afe78f29237e" (UID: "393a8d5d-f3b9-4846-ae16-afe78f29237e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.015108 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-scripts" (OuterVolumeSpecName: "scripts") pod "393a8d5d-f3b9-4846-ae16-afe78f29237e" (UID: "393a8d5d-f3b9-4846-ae16-afe78f29237e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.016140 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "393a8d5d-f3b9-4846-ae16-afe78f29237e" (UID: "393a8d5d-f3b9-4846-ae16-afe78f29237e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.116661 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.116733 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-config-data\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.116796 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.116820 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88c9m\" (UniqueName: \"kubernetes.io/projected/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-kube-api-access-88c9m\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.116853 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-scripts\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.116892 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.117082 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.117102 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.117115 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.117126 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/393a8d5d-f3b9-4846-ae16-afe78f29237e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.219322 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.219429 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-config-data\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.219895 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.219927 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88c9m\" (UniqueName: \"kubernetes.io/projected/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-kube-api-access-88c9m\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.219979 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-scripts\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.220019 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.220030 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.223853 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-config-data\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.224379 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.225115 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.225128 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-scripts\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.251359 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88c9m\" (UniqueName: \"kubernetes.io/projected/75efca1b-e080-4620-ad1b-8e1bcc3ba3ba-kube-api-access-88c9m\") pod \"manila-scheduler-0\" (UID: \"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba\") " pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.316967 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.831443 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 03 16:45:09 crc kubenswrapper[4744]: W1003 16:45:09.842690 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75efca1b_e080_4620_ad1b_8e1bcc3ba3ba.slice/crio-5233d6ba19a0da5b6b8d107360c5648f3d8ca1496f1f9455279784b91a710a71 WatchSource:0}: Error finding container 5233d6ba19a0da5b6b8d107360c5648f3d8ca1496f1f9455279784b91a710a71: Status 404 returned error can't find the container with id 5233d6ba19a0da5b6b8d107360c5648f3d8ca1496f1f9455279784b91a710a71 Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.923435 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:45:09 crc kubenswrapper[4744]: I1003 16:45:09.925393 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba","Type":"ContainerStarted","Data":"5233d6ba19a0da5b6b8d107360c5648f3d8ca1496f1f9455279784b91a710a71"} Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.046896 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.053575 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.082572 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.110117 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.111331 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.113568 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.114044 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.239658 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j7qv\" (UniqueName: \"kubernetes.io/projected/8fa06db6-8291-44c5-bd36-965c465ed970-kube-api-access-8j7qv\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.239712 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fa06db6-8291-44c5-bd36-965c465ed970-log-httpd\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.239742 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-config-data\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.239759 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fa06db6-8291-44c5-bd36-965c465ed970-run-httpd\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.239859 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.239932 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-scripts\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.239973 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.341377 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.342276 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j7qv\" (UniqueName: \"kubernetes.io/projected/8fa06db6-8291-44c5-bd36-965c465ed970-kube-api-access-8j7qv\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.342455 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fa06db6-8291-44c5-bd36-965c465ed970-log-httpd\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.342672 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-config-data\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.342789 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fa06db6-8291-44c5-bd36-965c465ed970-run-httpd\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.342934 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.343011 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fa06db6-8291-44c5-bd36-965c465ed970-log-httpd\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.343071 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fa06db6-8291-44c5-bd36-965c465ed970-run-httpd\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.343226 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-scripts\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.345964 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-scripts\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.346104 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.346375 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.360315 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-config-data\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.361649 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j7qv\" (UniqueName: \"kubernetes.io/projected/8fa06db6-8291-44c5-bd36-965c465ed970-kube-api-access-8j7qv\") pod \"ceilometer-0\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.438147 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.915903 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="393a8d5d-f3b9-4846-ae16-afe78f29237e" path="/var/lib/kubelet/pods/393a8d5d-f3b9-4846-ae16-afe78f29237e/volumes" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.917087 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5171a06a-cac6-4fa5-abfc-8423d21003ce" path="/var/lib/kubelet/pods/5171a06a-cac6-4fa5-abfc-8423d21003ce/volumes" Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.949360 4744 generic.go:334] "Generic (PLEG): container finished" podID="cb8803f7-c116-4175-b653-4b451e1da26d" containerID="44bc9bc5472efb50fcdd2da3dcd56942d91f2c82a55c8013f083377f384f0b06" exitCode=0 Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.949436 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb8803f7-c116-4175-b653-4b451e1da26d","Type":"ContainerDied","Data":"44bc9bc5472efb50fcdd2da3dcd56942d91f2c82a55c8013f083377f384f0b06"} Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.959872 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.963961 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba","Type":"ContainerStarted","Data":"0960b291397613f764688da12996be31b886833ec8a1e91a5709a824d6b3b741"} Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.964038 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"75efca1b-e080-4620-ad1b-8e1bcc3ba3ba","Type":"ContainerStarted","Data":"b4bdea65756e44b3b55a9c6cca388c29a7ca38a93103f827cb796c72ea29d868"} Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.978372 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 16:45:10 crc kubenswrapper[4744]: I1003 16:45:10.995371 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.995339821 podStartE2EDuration="2.995339821s" podCreationTimestamp="2025-10-03 16:45:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:45:10.98114788 +0000 UTC m=+1237.261023776" watchObservedRunningTime="2025-10-03 16:45:10.995339821 +0000 UTC m=+1237.275215717" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.037373 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.163426 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-config-data\") pod \"cb8803f7-c116-4175-b653-4b451e1da26d\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.163510 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb8803f7-c116-4175-b653-4b451e1da26d-httpd-run\") pod \"cb8803f7-c116-4175-b653-4b451e1da26d\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.163582 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-combined-ca-bundle\") pod \"cb8803f7-c116-4175-b653-4b451e1da26d\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.163614 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-internal-tls-certs\") pod \"cb8803f7-c116-4175-b653-4b451e1da26d\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.163649 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cb8803f7-c116-4175-b653-4b451e1da26d-ceph\") pod \"cb8803f7-c116-4175-b653-4b451e1da26d\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.163693 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsgnv\" (UniqueName: \"kubernetes.io/projected/cb8803f7-c116-4175-b653-4b451e1da26d-kube-api-access-lsgnv\") pod \"cb8803f7-c116-4175-b653-4b451e1da26d\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.163814 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-scripts\") pod \"cb8803f7-c116-4175-b653-4b451e1da26d\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.163851 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb8803f7-c116-4175-b653-4b451e1da26d-logs\") pod \"cb8803f7-c116-4175-b653-4b451e1da26d\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.163999 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"cb8803f7-c116-4175-b653-4b451e1da26d\" (UID: \"cb8803f7-c116-4175-b653-4b451e1da26d\") " Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.164024 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb8803f7-c116-4175-b653-4b451e1da26d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cb8803f7-c116-4175-b653-4b451e1da26d" (UID: "cb8803f7-c116-4175-b653-4b451e1da26d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.164511 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cb8803f7-c116-4175-b653-4b451e1da26d-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.164799 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb8803f7-c116-4175-b653-4b451e1da26d-logs" (OuterVolumeSpecName: "logs") pod "cb8803f7-c116-4175-b653-4b451e1da26d" (UID: "cb8803f7-c116-4175-b653-4b451e1da26d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.169365 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb8803f7-c116-4175-b653-4b451e1da26d-kube-api-access-lsgnv" (OuterVolumeSpecName: "kube-api-access-lsgnv") pod "cb8803f7-c116-4175-b653-4b451e1da26d" (UID: "cb8803f7-c116-4175-b653-4b451e1da26d"). InnerVolumeSpecName "kube-api-access-lsgnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.170198 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-scripts" (OuterVolumeSpecName: "scripts") pod "cb8803f7-c116-4175-b653-4b451e1da26d" (UID: "cb8803f7-c116-4175-b653-4b451e1da26d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.173660 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "cb8803f7-c116-4175-b653-4b451e1da26d" (UID: "cb8803f7-c116-4175-b653-4b451e1da26d"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.177686 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb8803f7-c116-4175-b653-4b451e1da26d-ceph" (OuterVolumeSpecName: "ceph") pod "cb8803f7-c116-4175-b653-4b451e1da26d" (UID: "cb8803f7-c116-4175-b653-4b451e1da26d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.206630 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb8803f7-c116-4175-b653-4b451e1da26d" (UID: "cb8803f7-c116-4175-b653-4b451e1da26d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.218718 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "cb8803f7-c116-4175-b653-4b451e1da26d" (UID: "cb8803f7-c116-4175-b653-4b451e1da26d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.225300 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-config-data" (OuterVolumeSpecName: "config-data") pod "cb8803f7-c116-4175-b653-4b451e1da26d" (UID: "cb8803f7-c116-4175-b653-4b451e1da26d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.266108 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.266153 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cb8803f7-c116-4175-b653-4b451e1da26d-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.266196 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.266211 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.266224 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.266241 4744 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb8803f7-c116-4175-b653-4b451e1da26d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.266255 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/cb8803f7-c116-4175-b653-4b451e1da26d-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.266266 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsgnv\" (UniqueName: \"kubernetes.io/projected/cb8803f7-c116-4175-b653-4b451e1da26d-kube-api-access-lsgnv\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.288870 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.368644 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.974284 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cb8803f7-c116-4175-b653-4b451e1da26d","Type":"ContainerDied","Data":"a392bc7af180766ab183d06d2805e964217faea000af24f9c19845fae717be91"} Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.974320 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.975172 4744 scope.go:117] "RemoveContainer" containerID="44bc9bc5472efb50fcdd2da3dcd56942d91f2c82a55c8013f083377f384f0b06" Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.977477 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fa06db6-8291-44c5-bd36-965c465ed970","Type":"ContainerStarted","Data":"ddd06e4de5ea54732fe7b9933e46b87deefcca81ee67200e32930825406411f5"} Oct 03 16:45:11 crc kubenswrapper[4744]: I1003 16:45:11.977593 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fa06db6-8291-44c5-bd36-965c465ed970","Type":"ContainerStarted","Data":"eb978b80fd3637f6e97b676c232f1f70cffc536d8f6b0f46517f96f0acb640f9"} Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.003405 4744 scope.go:117] "RemoveContainer" containerID="3c3fbe34545f10996e3b7700b35eb0c4743eb42323a910487b6f2b9afae26aaf" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.018050 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.043579 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.057957 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 16:45:12 crc kubenswrapper[4744]: E1003 16:45:12.058506 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb8803f7-c116-4175-b653-4b451e1da26d" containerName="glance-log" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.058525 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb8803f7-c116-4175-b653-4b451e1da26d" containerName="glance-log" Oct 03 16:45:12 crc kubenswrapper[4744]: E1003 16:45:12.058570 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb8803f7-c116-4175-b653-4b451e1da26d" containerName="glance-httpd" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.058579 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb8803f7-c116-4175-b653-4b451e1da26d" containerName="glance-httpd" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.058827 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb8803f7-c116-4175-b653-4b451e1da26d" containerName="glance-log" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.058847 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb8803f7-c116-4175-b653-4b451e1da26d" containerName="glance-httpd" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.060112 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.074056 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.074428 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.091370 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.193646 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.193724 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.193755 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.193775 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.193790 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.194009 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqdbm\" (UniqueName: \"kubernetes.io/projected/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-kube-api-access-rqdbm\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.194107 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.194250 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-logs\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.194310 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.295805 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqdbm\" (UniqueName: \"kubernetes.io/projected/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-kube-api-access-rqdbm\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.295856 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.295907 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-logs\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.295928 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.295969 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.296011 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.296036 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.296056 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.296071 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.296374 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.296442 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-logs\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.296518 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.300415 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.302473 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.303020 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.304153 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.305154 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.320271 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqdbm\" (UniqueName: \"kubernetes.io/projected/b7679ad4-70b7-4058-a2e6-c4dba8a409e9-kube-api-access-rqdbm\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.326373 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"b7679ad4-70b7-4058-a2e6-c4dba8a409e9\") " pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.411065 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.910662 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb8803f7-c116-4175-b653-4b451e1da26d" path="/var/lib/kubelet/pods/cb8803f7-c116-4175-b653-4b451e1da26d/volumes" Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.990617 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fa06db6-8291-44c5-bd36-965c465ed970","Type":"ContainerStarted","Data":"2be38fa7c5539b3eb33d6978b3f255b4fdaecebc80acb76b01c5f42261357222"} Oct 03 16:45:12 crc kubenswrapper[4744]: I1003 16:45:12.992024 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 16:45:14 crc kubenswrapper[4744]: I1003 16:45:14.015437 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b7679ad4-70b7-4058-a2e6-c4dba8a409e9","Type":"ContainerStarted","Data":"9ae9d76c7386ab5eeb2ab2529c299f84141baf78b0cb4170db5d6755662571b4"} Oct 03 16:45:14 crc kubenswrapper[4744]: I1003 16:45:14.015903 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b7679ad4-70b7-4058-a2e6-c4dba8a409e9","Type":"ContainerStarted","Data":"7b4724ed3d4ca9b1946a26fb15578f1200e0b50f9ccca073ac831b35abdd76e4"} Oct 03 16:45:14 crc kubenswrapper[4744]: I1003 16:45:14.020585 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fa06db6-8291-44c5-bd36-965c465ed970","Type":"ContainerStarted","Data":"75122b72b34cab11d612bbe635e1de79825bf14c17fa5a4800fc8a9e02ca265c"} Oct 03 16:45:14 crc kubenswrapper[4744]: I1003 16:45:14.997834 4744 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod102bbbfd-8dd8-4946-9f35-a22a87ebaaed"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod102bbbfd-8dd8-4946-9f35-a22a87ebaaed] : Timed out while waiting for systemd to remove kubepods-besteffort-pod102bbbfd_8dd8_4946_9f35_a22a87ebaaed.slice" Oct 03 16:45:14 crc kubenswrapper[4744]: E1003 16:45:14.998391 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod102bbbfd-8dd8-4946-9f35-a22a87ebaaed] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod102bbbfd-8dd8-4946-9f35-a22a87ebaaed] : Timed out while waiting for systemd to remove kubepods-besteffort-pod102bbbfd_8dd8_4946_9f35_a22a87ebaaed.slice" pod="openstack/barbican-api-5467484d76-wdx4m" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.030911 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b7679ad4-70b7-4058-a2e6-c4dba8a409e9","Type":"ContainerStarted","Data":"e27eace46d7881125c8d112aa1ee454c5dafd1618947ef37617954814caec9e4"} Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.034957 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5467484d76-wdx4m" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.035096 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fa06db6-8291-44c5-bd36-965c465ed970","Type":"ContainerStarted","Data":"9407e0b7e8a5a7381cdcd46fca97ebd02c2143d6399cf5c16b4614d9dc90aedf"} Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.035711 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.055476 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.055453829 podStartE2EDuration="3.055453829s" podCreationTimestamp="2025-10-03 16:45:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:45:15.051213661 +0000 UTC m=+1241.331089577" watchObservedRunningTime="2025-10-03 16:45:15.055453829 +0000 UTC m=+1241.335329725" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.073902 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5467484d76-wdx4m"] Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.085431 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5467484d76-wdx4m"] Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.087971 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.4586008879999999 podStartE2EDuration="5.087960045s" podCreationTimestamp="2025-10-03 16:45:10 +0000 UTC" firstStartedPulling="2025-10-03 16:45:10.977308882 +0000 UTC m=+1237.257184778" lastFinishedPulling="2025-10-03 16:45:14.606668049 +0000 UTC m=+1240.886543935" observedRunningTime="2025-10-03 16:45:15.081656245 +0000 UTC m=+1241.361532171" watchObservedRunningTime="2025-10-03 16:45:15.087960045 +0000 UTC m=+1241.367835941" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.116345 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.194127 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.247486 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-7097-account-create-kzss9"] Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.249839 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7097-account-create-kzss9" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.256686 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.263410 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7097-account-create-kzss9"] Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.363691 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79686\" (UniqueName: \"kubernetes.io/projected/fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5-kube-api-access-79686\") pod \"nova-cell0-7097-account-create-kzss9\" (UID: \"fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5\") " pod="openstack/nova-cell0-7097-account-create-kzss9" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.455623 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-fc01-account-create-5cmg6"] Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.456796 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-fc01-account-create-5cmg6" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.458761 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.465613 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79686\" (UniqueName: \"kubernetes.io/projected/fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5-kube-api-access-79686\") pod \"nova-cell0-7097-account-create-kzss9\" (UID: \"fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5\") " pod="openstack/nova-cell0-7097-account-create-kzss9" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.468669 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-fc01-account-create-5cmg6"] Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.513706 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79686\" (UniqueName: \"kubernetes.io/projected/fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5-kube-api-access-79686\") pod \"nova-cell0-7097-account-create-kzss9\" (UID: \"fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5\") " pod="openstack/nova-cell0-7097-account-create-kzss9" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.568001 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwr6n\" (UniqueName: \"kubernetes.io/projected/beff32b5-71c6-4be0-8a7f-5abae594db52-kube-api-access-rwr6n\") pod \"nova-cell1-fc01-account-create-5cmg6\" (UID: \"beff32b5-71c6-4be0-8a7f-5abae594db52\") " pod="openstack/nova-cell1-fc01-account-create-5cmg6" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.574919 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7097-account-create-kzss9" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.670369 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwr6n\" (UniqueName: \"kubernetes.io/projected/beff32b5-71c6-4be0-8a7f-5abae594db52-kube-api-access-rwr6n\") pod \"nova-cell1-fc01-account-create-5cmg6\" (UID: \"beff32b5-71c6-4be0-8a7f-5abae594db52\") " pod="openstack/nova-cell1-fc01-account-create-5cmg6" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.697849 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwr6n\" (UniqueName: \"kubernetes.io/projected/beff32b5-71c6-4be0-8a7f-5abae594db52-kube-api-access-rwr6n\") pod \"nova-cell1-fc01-account-create-5cmg6\" (UID: \"beff32b5-71c6-4be0-8a7f-5abae594db52\") " pod="openstack/nova-cell1-fc01-account-create-5cmg6" Oct 03 16:45:15 crc kubenswrapper[4744]: I1003 16:45:15.786334 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-fc01-account-create-5cmg6" Oct 03 16:45:16 crc kubenswrapper[4744]: I1003 16:45:16.044255 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="4fa451ca-3be8-4074-b903-5ea114f47bda" containerName="manila-share" containerID="cri-o://6292a918a1fbf56b48a41e7b1d948dba59c8a26c8e38e368dfb1779a3c68d8f2" gracePeriod=30 Oct 03 16:45:16 crc kubenswrapper[4744]: I1003 16:45:16.044527 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="4fa451ca-3be8-4074-b903-5ea114f47bda" containerName="probe" containerID="cri-o://53ac3655289c7648b5a9b1b7f77edd67983324dc05b6862f0ec695326ada3bb9" gracePeriod=30 Oct 03 16:45:16 crc kubenswrapper[4744]: I1003 16:45:16.045683 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7097-account-create-kzss9"] Oct 03 16:45:16 crc kubenswrapper[4744]: I1003 16:45:16.218853 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-fc01-account-create-5cmg6"] Oct 03 16:45:16 crc kubenswrapper[4744]: W1003 16:45:16.220154 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbeff32b5_71c6_4be0_8a7f_5abae594db52.slice/crio-24797ad5dbcf9daf17001ba7b2ac34d7ba09b234467463b21b222828d444427f WatchSource:0}: Error finding container 24797ad5dbcf9daf17001ba7b2ac34d7ba09b234467463b21b222828d444427f: Status 404 returned error can't find the container with id 24797ad5dbcf9daf17001ba7b2ac34d7ba09b234467463b21b222828d444427f Oct 03 16:45:16 crc kubenswrapper[4744]: I1003 16:45:16.887699 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 16:45:16 crc kubenswrapper[4744]: I1003 16:45:16.888161 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f524b0a3-3650-4f2e-9a27-d74b82e3e06d" containerName="glance-log" containerID="cri-o://502896ddc00c75f7a96640a4e1f09fb28ca1e152a7bd09d0d12cd4ecdf2ac7be" gracePeriod=30 Oct 03 16:45:16 crc kubenswrapper[4744]: I1003 16:45:16.888294 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f524b0a3-3650-4f2e-9a27-d74b82e3e06d" containerName="glance-httpd" containerID="cri-o://85cb0fee8dd78b21ef8c932996f2e53b5687f54c2bceb977c555d98b63cba81d" gracePeriod=30 Oct 03 16:45:16 crc kubenswrapper[4744]: I1003 16:45:16.974210 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="102bbbfd-8dd8-4946-9f35-a22a87ebaaed" path="/var/lib/kubelet/pods/102bbbfd-8dd8-4946-9f35-a22a87ebaaed/volumes" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.053626 4744 generic.go:334] "Generic (PLEG): container finished" podID="beff32b5-71c6-4be0-8a7f-5abae594db52" containerID="f5a1edcf8dc2bd8f953414f1291177ce5631944d6d835b74686c4d7230d511fa" exitCode=0 Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.053687 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-fc01-account-create-5cmg6" event={"ID":"beff32b5-71c6-4be0-8a7f-5abae594db52","Type":"ContainerDied","Data":"f5a1edcf8dc2bd8f953414f1291177ce5631944d6d835b74686c4d7230d511fa"} Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.053711 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-fc01-account-create-5cmg6" event={"ID":"beff32b5-71c6-4be0-8a7f-5abae594db52","Type":"ContainerStarted","Data":"24797ad5dbcf9daf17001ba7b2ac34d7ba09b234467463b21b222828d444427f"} Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.057532 4744 generic.go:334] "Generic (PLEG): container finished" podID="4fa451ca-3be8-4074-b903-5ea114f47bda" containerID="53ac3655289c7648b5a9b1b7f77edd67983324dc05b6862f0ec695326ada3bb9" exitCode=0 Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.057572 4744 generic.go:334] "Generic (PLEG): container finished" podID="4fa451ca-3be8-4074-b903-5ea114f47bda" containerID="6292a918a1fbf56b48a41e7b1d948dba59c8a26c8e38e368dfb1779a3c68d8f2" exitCode=1 Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.057621 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"4fa451ca-3be8-4074-b903-5ea114f47bda","Type":"ContainerDied","Data":"53ac3655289c7648b5a9b1b7f77edd67983324dc05b6862f0ec695326ada3bb9"} Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.057653 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"4fa451ca-3be8-4074-b903-5ea114f47bda","Type":"ContainerDied","Data":"6292a918a1fbf56b48a41e7b1d948dba59c8a26c8e38e368dfb1779a3c68d8f2"} Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.057664 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"4fa451ca-3be8-4074-b903-5ea114f47bda","Type":"ContainerDied","Data":"fa4c3c41d905a5c33b1f8bc53a6bab59c974a52e53b72f7b6ffface0dcbd30d0"} Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.057674 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa4c3c41d905a5c33b1f8bc53a6bab59c974a52e53b72f7b6ffface0dcbd30d0" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.059790 4744 generic.go:334] "Generic (PLEG): container finished" podID="fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5" containerID="bafe7e2bfc3a5219bf6833e1abe232b87d2682a3d9b40fadd8034609386721da" exitCode=0 Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.059879 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7097-account-create-kzss9" event={"ID":"fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5","Type":"ContainerDied","Data":"bafe7e2bfc3a5219bf6833e1abe232b87d2682a3d9b40fadd8034609386721da"} Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.059899 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7097-account-create-kzss9" event={"ID":"fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5","Type":"ContainerStarted","Data":"c53961df5a3c2757d13a1f557df80853681838743e1d2020ea26ace0401a8109"} Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.061956 4744 generic.go:334] "Generic (PLEG): container finished" podID="f524b0a3-3650-4f2e-9a27-d74b82e3e06d" containerID="502896ddc00c75f7a96640a4e1f09fb28ca1e152a7bd09d0d12cd4ecdf2ac7be" exitCode=143 Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.061989 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f524b0a3-3650-4f2e-9a27-d74b82e3e06d","Type":"ContainerDied","Data":"502896ddc00c75f7a96640a4e1f09fb28ca1e152a7bd09d0d12cd4ecdf2ac7be"} Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.111571 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.208363 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-combined-ca-bundle\") pod \"4fa451ca-3be8-4074-b903-5ea114f47bda\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.208427 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ld92\" (UniqueName: \"kubernetes.io/projected/4fa451ca-3be8-4074-b903-5ea114f47bda-kube-api-access-9ld92\") pod \"4fa451ca-3be8-4074-b903-5ea114f47bda\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.208482 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/4fa451ca-3be8-4074-b903-5ea114f47bda-var-lib-manila\") pod \"4fa451ca-3be8-4074-b903-5ea114f47bda\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.208547 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-scripts\") pod \"4fa451ca-3be8-4074-b903-5ea114f47bda\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.208603 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-config-data-custom\") pod \"4fa451ca-3be8-4074-b903-5ea114f47bda\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.208658 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-config-data\") pod \"4fa451ca-3be8-4074-b903-5ea114f47bda\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.208716 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4fa451ca-3be8-4074-b903-5ea114f47bda-etc-machine-id\") pod \"4fa451ca-3be8-4074-b903-5ea114f47bda\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.208767 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4fa451ca-3be8-4074-b903-5ea114f47bda-ceph\") pod \"4fa451ca-3be8-4074-b903-5ea114f47bda\" (UID: \"4fa451ca-3be8-4074-b903-5ea114f47bda\") " Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.209828 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4fa451ca-3be8-4074-b903-5ea114f47bda-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "4fa451ca-3be8-4074-b903-5ea114f47bda" (UID: "4fa451ca-3be8-4074-b903-5ea114f47bda"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.209977 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4fa451ca-3be8-4074-b903-5ea114f47bda-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4fa451ca-3be8-4074-b903-5ea114f47bda" (UID: "4fa451ca-3be8-4074-b903-5ea114f47bda"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.215995 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fa451ca-3be8-4074-b903-5ea114f47bda-kube-api-access-9ld92" (OuterVolumeSpecName: "kube-api-access-9ld92") pod "4fa451ca-3be8-4074-b903-5ea114f47bda" (UID: "4fa451ca-3be8-4074-b903-5ea114f47bda"). InnerVolumeSpecName "kube-api-access-9ld92". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.221747 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4fa451ca-3be8-4074-b903-5ea114f47bda" (UID: "4fa451ca-3be8-4074-b903-5ea114f47bda"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.228128 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-scripts" (OuterVolumeSpecName: "scripts") pod "4fa451ca-3be8-4074-b903-5ea114f47bda" (UID: "4fa451ca-3be8-4074-b903-5ea114f47bda"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.235976 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fa451ca-3be8-4074-b903-5ea114f47bda-ceph" (OuterVolumeSpecName: "ceph") pod "4fa451ca-3be8-4074-b903-5ea114f47bda" (UID: "4fa451ca-3be8-4074-b903-5ea114f47bda"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.279586 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4fa451ca-3be8-4074-b903-5ea114f47bda" (UID: "4fa451ca-3be8-4074-b903-5ea114f47bda"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.311384 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.311419 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ld92\" (UniqueName: \"kubernetes.io/projected/4fa451ca-3be8-4074-b903-5ea114f47bda-kube-api-access-9ld92\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.311433 4744 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/4fa451ca-3be8-4074-b903-5ea114f47bda-var-lib-manila\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.311448 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.311460 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.311468 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4fa451ca-3be8-4074-b903-5ea114f47bda-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.311479 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/4fa451ca-3be8-4074-b903-5ea114f47bda-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.353981 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-config-data" (OuterVolumeSpecName: "config-data") pod "4fa451ca-3be8-4074-b903-5ea114f47bda" (UID: "4fa451ca-3be8-4074-b903-5ea114f47bda"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:17 crc kubenswrapper[4744]: I1003 16:45:17.412939 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fa451ca-3be8-4074-b903-5ea114f47bda-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.069480 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.116712 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.132484 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.141912 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 03 16:45:18 crc kubenswrapper[4744]: E1003 16:45:18.142656 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fa451ca-3be8-4074-b903-5ea114f47bda" containerName="probe" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.142739 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fa451ca-3be8-4074-b903-5ea114f47bda" containerName="probe" Oct 03 16:45:18 crc kubenswrapper[4744]: E1003 16:45:18.142804 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fa451ca-3be8-4074-b903-5ea114f47bda" containerName="manila-share" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.142857 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fa451ca-3be8-4074-b903-5ea114f47bda" containerName="manila-share" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.143064 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fa451ca-3be8-4074-b903-5ea114f47bda" containerName="probe" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.143146 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fa451ca-3be8-4074-b903-5ea114f47bda" containerName="manila-share" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.144148 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.147092 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.150844 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.233861 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3c066c82-d9c3-40bc-805d-16a45e328e49-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.233916 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3c066c82-d9c3-40bc-805d-16a45e328e49-ceph\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.233970 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c066c82-d9c3-40bc-805d-16a45e328e49-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.234014 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2xm5\" (UniqueName: \"kubernetes.io/projected/3c066c82-d9c3-40bc-805d-16a45e328e49-kube-api-access-b2xm5\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.234043 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c066c82-d9c3-40bc-805d-16a45e328e49-scripts\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.234099 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c066c82-d9c3-40bc-805d-16a45e328e49-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.234117 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c066c82-d9c3-40bc-805d-16a45e328e49-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.234134 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c066c82-d9c3-40bc-805d-16a45e328e49-config-data\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.336264 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3c066c82-d9c3-40bc-805d-16a45e328e49-ceph\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.336397 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c066c82-d9c3-40bc-805d-16a45e328e49-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.336483 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2xm5\" (UniqueName: \"kubernetes.io/projected/3c066c82-d9c3-40bc-805d-16a45e328e49-kube-api-access-b2xm5\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.336549 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c066c82-d9c3-40bc-805d-16a45e328e49-scripts\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.336637 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c066c82-d9c3-40bc-805d-16a45e328e49-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.336688 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c066c82-d9c3-40bc-805d-16a45e328e49-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.336715 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c066c82-d9c3-40bc-805d-16a45e328e49-config-data\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.336799 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3c066c82-d9c3-40bc-805d-16a45e328e49-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.336968 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/3c066c82-d9c3-40bc-805d-16a45e328e49-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.338853 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3c066c82-d9c3-40bc-805d-16a45e328e49-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.342729 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c066c82-d9c3-40bc-805d-16a45e328e49-config-data\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.343314 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c066c82-d9c3-40bc-805d-16a45e328e49-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.346259 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c066c82-d9c3-40bc-805d-16a45e328e49-scripts\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.352171 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3c066c82-d9c3-40bc-805d-16a45e328e49-ceph\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.376236 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c066c82-d9c3-40bc-805d-16a45e328e49-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.377713 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2xm5\" (UniqueName: \"kubernetes.io/projected/3c066c82-d9c3-40bc-805d-16a45e328e49-kube-api-access-b2xm5\") pod \"manila-share-share1-0\" (UID: \"3c066c82-d9c3-40bc-805d-16a45e328e49\") " pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.469088 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.545999 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-fc01-account-create-5cmg6" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.560410 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7097-account-create-kzss9" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.643699 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwr6n\" (UniqueName: \"kubernetes.io/projected/beff32b5-71c6-4be0-8a7f-5abae594db52-kube-api-access-rwr6n\") pod \"beff32b5-71c6-4be0-8a7f-5abae594db52\" (UID: \"beff32b5-71c6-4be0-8a7f-5abae594db52\") " Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.643836 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79686\" (UniqueName: \"kubernetes.io/projected/fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5-kube-api-access-79686\") pod \"fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5\" (UID: \"fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5\") " Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.648852 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5-kube-api-access-79686" (OuterVolumeSpecName: "kube-api-access-79686") pod "fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5" (UID: "fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5"). InnerVolumeSpecName "kube-api-access-79686". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.671213 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beff32b5-71c6-4be0-8a7f-5abae594db52-kube-api-access-rwr6n" (OuterVolumeSpecName: "kube-api-access-rwr6n") pod "beff32b5-71c6-4be0-8a7f-5abae594db52" (UID: "beff32b5-71c6-4be0-8a7f-5abae594db52"). InnerVolumeSpecName "kube-api-access-rwr6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.746400 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79686\" (UniqueName: \"kubernetes.io/projected/fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5-kube-api-access-79686\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.746451 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwr6n\" (UniqueName: \"kubernetes.io/projected/beff32b5-71c6-4be0-8a7f-5abae594db52-kube-api-access-rwr6n\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:18 crc kubenswrapper[4744]: I1003 16:45:18.903420 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fa451ca-3be8-4074-b903-5ea114f47bda" path="/var/lib/kubelet/pods/4fa451ca-3be8-4074-b903-5ea114f47bda/volumes" Oct 03 16:45:19 crc kubenswrapper[4744]: I1003 16:45:19.025003 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 03 16:45:19 crc kubenswrapper[4744]: W1003 16:45:19.029969 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c066c82_d9c3_40bc_805d_16a45e328e49.slice/crio-be7220fa9b59f4ba94763b5b289d42c8e54984c4ebafed5d06fac053efed7df9 WatchSource:0}: Error finding container be7220fa9b59f4ba94763b5b289d42c8e54984c4ebafed5d06fac053efed7df9: Status 404 returned error can't find the container with id be7220fa9b59f4ba94763b5b289d42c8e54984c4ebafed5d06fac053efed7df9 Oct 03 16:45:19 crc kubenswrapper[4744]: I1003 16:45:19.080415 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-fc01-account-create-5cmg6" event={"ID":"beff32b5-71c6-4be0-8a7f-5abae594db52","Type":"ContainerDied","Data":"24797ad5dbcf9daf17001ba7b2ac34d7ba09b234467463b21b222828d444427f"} Oct 03 16:45:19 crc kubenswrapper[4744]: I1003 16:45:19.080464 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24797ad5dbcf9daf17001ba7b2ac34d7ba09b234467463b21b222828d444427f" Oct 03 16:45:19 crc kubenswrapper[4744]: I1003 16:45:19.080571 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-fc01-account-create-5cmg6" Oct 03 16:45:19 crc kubenswrapper[4744]: I1003 16:45:19.082600 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7097-account-create-kzss9" event={"ID":"fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5","Type":"ContainerDied","Data":"c53961df5a3c2757d13a1f557df80853681838743e1d2020ea26ace0401a8109"} Oct 03 16:45:19 crc kubenswrapper[4744]: I1003 16:45:19.082637 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c53961df5a3c2757d13a1f557df80853681838743e1d2020ea26ace0401a8109" Oct 03 16:45:19 crc kubenswrapper[4744]: I1003 16:45:19.082693 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7097-account-create-kzss9" Oct 03 16:45:19 crc kubenswrapper[4744]: I1003 16:45:19.084842 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3c066c82-d9c3-40bc-805d-16a45e328e49","Type":"ContainerStarted","Data":"be7220fa9b59f4ba94763b5b289d42c8e54984c4ebafed5d06fac053efed7df9"} Oct 03 16:45:19 crc kubenswrapper[4744]: I1003 16:45:19.317631 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.095341 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3c066c82-d9c3-40bc-805d-16a45e328e49","Type":"ContainerStarted","Data":"5034ea55d3f497a4bfea4acf6d0135c2adc20439f627d94b018ad42d30fe13c9"} Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.095722 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"3c066c82-d9c3-40bc-805d-16a45e328e49","Type":"ContainerStarted","Data":"e3b314af70c5dda4f4f219af4b18e54bfcbbd0f6eb07cd51c4b6d6055835d2e4"} Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.153771 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.15374448 podStartE2EDuration="2.15374448s" podCreationTimestamp="2025-10-03 16:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:45:20.148406454 +0000 UTC m=+1246.428282350" watchObservedRunningTime="2025-10-03 16:45:20.15374448 +0000 UTC m=+1246.433620376" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.636566 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zgrsm"] Oct 03 16:45:20 crc kubenswrapper[4744]: E1003 16:45:20.636936 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beff32b5-71c6-4be0-8a7f-5abae594db52" containerName="mariadb-account-create" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.636952 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="beff32b5-71c6-4be0-8a7f-5abae594db52" containerName="mariadb-account-create" Oct 03 16:45:20 crc kubenswrapper[4744]: E1003 16:45:20.636992 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5" containerName="mariadb-account-create" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.636999 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5" containerName="mariadb-account-create" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.637149 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="beff32b5-71c6-4be0-8a7f-5abae594db52" containerName="mariadb-account-create" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.637167 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5" containerName="mariadb-account-create" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.637763 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.640027 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-p2fmp" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.640388 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.640584 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.654774 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zgrsm"] Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.747758 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.792929 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsmvm\" (UniqueName: \"kubernetes.io/projected/1cd09fa9-4967-49a1-a086-55405a0a681b-kube-api-access-fsmvm\") pod \"nova-cell0-conductor-db-sync-zgrsm\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.793025 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-zgrsm\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.793063 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-scripts\") pod \"nova-cell0-conductor-db-sync-zgrsm\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.793189 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-config-data\") pod \"nova-cell0-conductor-db-sync-zgrsm\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.894576 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-scripts\") pod \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.894652 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-httpd-run\") pod \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.894699 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-combined-ca-bundle\") pod \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.894758 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-config-data\") pod \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.894800 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwdpp\" (UniqueName: \"kubernetes.io/projected/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-kube-api-access-xwdpp\") pod \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.894894 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-ceph\") pod \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.894917 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-logs\") pod \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.895001 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.895034 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-public-tls-certs\") pod \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\" (UID: \"f524b0a3-3650-4f2e-9a27-d74b82e3e06d\") " Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.895240 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f524b0a3-3650-4f2e-9a27-d74b82e3e06d" (UID: "f524b0a3-3650-4f2e-9a27-d74b82e3e06d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.895371 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-zgrsm\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.895432 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-scripts\") pod \"nova-cell0-conductor-db-sync-zgrsm\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.895567 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-config-data\") pod \"nova-cell0-conductor-db-sync-zgrsm\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.895639 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsmvm\" (UniqueName: \"kubernetes.io/projected/1cd09fa9-4967-49a1-a086-55405a0a681b-kube-api-access-fsmvm\") pod \"nova-cell0-conductor-db-sync-zgrsm\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.895718 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.897611 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-logs" (OuterVolumeSpecName: "logs") pod "f524b0a3-3650-4f2e-9a27-d74b82e3e06d" (UID: "f524b0a3-3650-4f2e-9a27-d74b82e3e06d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.904997 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-config-data\") pod \"nova-cell0-conductor-db-sync-zgrsm\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.905904 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-zgrsm\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.911041 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-kube-api-access-xwdpp" (OuterVolumeSpecName: "kube-api-access-xwdpp") pod "f524b0a3-3650-4f2e-9a27-d74b82e3e06d" (UID: "f524b0a3-3650-4f2e-9a27-d74b82e3e06d"). InnerVolumeSpecName "kube-api-access-xwdpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.911464 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsmvm\" (UniqueName: \"kubernetes.io/projected/1cd09fa9-4967-49a1-a086-55405a0a681b-kube-api-access-fsmvm\") pod \"nova-cell0-conductor-db-sync-zgrsm\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.911632 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-scripts" (OuterVolumeSpecName: "scripts") pod "f524b0a3-3650-4f2e-9a27-d74b82e3e06d" (UID: "f524b0a3-3650-4f2e-9a27-d74b82e3e06d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.925487 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-scripts\") pod \"nova-cell0-conductor-db-sync-zgrsm\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.926270 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-ceph" (OuterVolumeSpecName: "ceph") pod "f524b0a3-3650-4f2e-9a27-d74b82e3e06d" (UID: "f524b0a3-3650-4f2e-9a27-d74b82e3e06d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.927138 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f524b0a3-3650-4f2e-9a27-d74b82e3e06d" (UID: "f524b0a3-3650-4f2e-9a27-d74b82e3e06d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.930689 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "f524b0a3-3650-4f2e-9a27-d74b82e3e06d" (UID: "f524b0a3-3650-4f2e-9a27-d74b82e3e06d"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.951793 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f524b0a3-3650-4f2e-9a27-d74b82e3e06d" (UID: "f524b0a3-3650-4f2e-9a27-d74b82e3e06d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:20 crc kubenswrapper[4744]: I1003 16:45:20.968008 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-config-data" (OuterVolumeSpecName: "config-data") pod "f524b0a3-3650-4f2e-9a27-d74b82e3e06d" (UID: "f524b0a3-3650-4f2e-9a27-d74b82e3e06d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.000040 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.000291 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.000356 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.000411 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwdpp\" (UniqueName: \"kubernetes.io/projected/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-kube-api-access-xwdpp\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.000479 4744 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-ceph\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.000582 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.000702 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.000799 4744 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f524b0a3-3650-4f2e-9a27-d74b82e3e06d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.022736 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.060648 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.103005 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.113319 4744 generic.go:334] "Generic (PLEG): container finished" podID="f524b0a3-3650-4f2e-9a27-d74b82e3e06d" containerID="85cb0fee8dd78b21ef8c932996f2e53b5687f54c2bceb977c555d98b63cba81d" exitCode=0 Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.113677 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.113660 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f524b0a3-3650-4f2e-9a27-d74b82e3e06d","Type":"ContainerDied","Data":"85cb0fee8dd78b21ef8c932996f2e53b5687f54c2bceb977c555d98b63cba81d"} Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.113900 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f524b0a3-3650-4f2e-9a27-d74b82e3e06d","Type":"ContainerDied","Data":"31a435ffe66addb6d519f50e0249a245a72d47ac07255246f9eb9501ff114a43"} Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.113923 4744 scope.go:117] "RemoveContainer" containerID="85cb0fee8dd78b21ef8c932996f2e53b5687f54c2bceb977c555d98b63cba81d" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.149476 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.151729 4744 scope.go:117] "RemoveContainer" containerID="502896ddc00c75f7a96640a4e1f09fb28ca1e152a7bd09d0d12cd4ecdf2ac7be" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.164252 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.192292 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 16:45:21 crc kubenswrapper[4744]: E1003 16:45:21.193214 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f524b0a3-3650-4f2e-9a27-d74b82e3e06d" containerName="glance-httpd" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.193355 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f524b0a3-3650-4f2e-9a27-d74b82e3e06d" containerName="glance-httpd" Oct 03 16:45:21 crc kubenswrapper[4744]: E1003 16:45:21.193464 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f524b0a3-3650-4f2e-9a27-d74b82e3e06d" containerName="glance-log" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.193610 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f524b0a3-3650-4f2e-9a27-d74b82e3e06d" containerName="glance-log" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.193946 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f524b0a3-3650-4f2e-9a27-d74b82e3e06d" containerName="glance-log" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.194043 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f524b0a3-3650-4f2e-9a27-d74b82e3e06d" containerName="glance-httpd" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.195372 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.198200 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.201264 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.214334 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.237661 4744 scope.go:117] "RemoveContainer" containerID="85cb0fee8dd78b21ef8c932996f2e53b5687f54c2bceb977c555d98b63cba81d" Oct 03 16:45:21 crc kubenswrapper[4744]: E1003 16:45:21.238518 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85cb0fee8dd78b21ef8c932996f2e53b5687f54c2bceb977c555d98b63cba81d\": container with ID starting with 85cb0fee8dd78b21ef8c932996f2e53b5687f54c2bceb977c555d98b63cba81d not found: ID does not exist" containerID="85cb0fee8dd78b21ef8c932996f2e53b5687f54c2bceb977c555d98b63cba81d" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.238569 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85cb0fee8dd78b21ef8c932996f2e53b5687f54c2bceb977c555d98b63cba81d"} err="failed to get container status \"85cb0fee8dd78b21ef8c932996f2e53b5687f54c2bceb977c555d98b63cba81d\": rpc error: code = NotFound desc = could not find container \"85cb0fee8dd78b21ef8c932996f2e53b5687f54c2bceb977c555d98b63cba81d\": container with ID starting with 85cb0fee8dd78b21ef8c932996f2e53b5687f54c2bceb977c555d98b63cba81d not found: ID does not exist" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.238600 4744 scope.go:117] "RemoveContainer" containerID="502896ddc00c75f7a96640a4e1f09fb28ca1e152a7bd09d0d12cd4ecdf2ac7be" Oct 03 16:45:21 crc kubenswrapper[4744]: E1003 16:45:21.239092 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"502896ddc00c75f7a96640a4e1f09fb28ca1e152a7bd09d0d12cd4ecdf2ac7be\": container with ID starting with 502896ddc00c75f7a96640a4e1f09fb28ca1e152a7bd09d0d12cd4ecdf2ac7be not found: ID does not exist" containerID="502896ddc00c75f7a96640a4e1f09fb28ca1e152a7bd09d0d12cd4ecdf2ac7be" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.239125 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"502896ddc00c75f7a96640a4e1f09fb28ca1e152a7bd09d0d12cd4ecdf2ac7be"} err="failed to get container status \"502896ddc00c75f7a96640a4e1f09fb28ca1e152a7bd09d0d12cd4ecdf2ac7be\": rpc error: code = NotFound desc = could not find container \"502896ddc00c75f7a96640a4e1f09fb28ca1e152a7bd09d0d12cd4ecdf2ac7be\": container with ID starting with 502896ddc00c75f7a96640a4e1f09fb28ca1e152a7bd09d0d12cd4ecdf2ac7be not found: ID does not exist" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.308446 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.308530 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-logs\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.308551 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jptd\" (UniqueName: \"kubernetes.io/projected/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-kube-api-access-2jptd\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.308681 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-config-data\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.308724 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-scripts\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.308740 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.308776 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.308798 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.308814 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-ceph\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.410682 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-logs\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.410742 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jptd\" (UniqueName: \"kubernetes.io/projected/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-kube-api-access-2jptd\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.410790 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-config-data\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.410827 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-scripts\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.410842 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.410876 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.410897 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.410914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-ceph\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.410983 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.411244 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.411783 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.412160 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-logs\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.432460 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.433136 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-scripts\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.433597 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-ceph\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.433972 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-config-data\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.439553 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jptd\" (UniqueName: \"kubernetes.io/projected/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-kube-api-access-2jptd\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.439805 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/411dd208-5058-4f51-b5ab-4e4f8c77cd8e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.465834 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"411dd208-5058-4f51-b5ab-4e4f8c77cd8e\") " pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.515213 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.554432 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zgrsm"] Oct 03 16:45:21 crc kubenswrapper[4744]: W1003 16:45:21.562154 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1cd09fa9_4967_49a1_a086_55405a0a681b.slice/crio-645a07692dd7b8d15c29d16d28da0c05a077e77fafb0ebba682f1f84898ed1e2 WatchSource:0}: Error finding container 645a07692dd7b8d15c29d16d28da0c05a077e77fafb0ebba682f1f84898ed1e2: Status 404 returned error can't find the container with id 645a07692dd7b8d15c29d16d28da0c05a077e77fafb0ebba682f1f84898ed1e2 Oct 03 16:45:21 crc kubenswrapper[4744]: I1003 16:45:21.979727 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 03 16:45:22 crc kubenswrapper[4744]: I1003 16:45:22.133870 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zgrsm" event={"ID":"1cd09fa9-4967-49a1-a086-55405a0a681b","Type":"ContainerStarted","Data":"645a07692dd7b8d15c29d16d28da0c05a077e77fafb0ebba682f1f84898ed1e2"} Oct 03 16:45:22 crc kubenswrapper[4744]: I1003 16:45:22.137673 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 16:45:22 crc kubenswrapper[4744]: W1003 16:45:22.152243 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod411dd208_5058_4f51_b5ab_4e4f8c77cd8e.slice/crio-f93584cc754490792bd18f4cf0b75deb30a5568fbf615e0596a8169c7a450f28 WatchSource:0}: Error finding container f93584cc754490792bd18f4cf0b75deb30a5568fbf615e0596a8169c7a450f28: Status 404 returned error can't find the container with id f93584cc754490792bd18f4cf0b75deb30a5568fbf615e0596a8169c7a450f28 Oct 03 16:45:22 crc kubenswrapper[4744]: I1003 16:45:22.411899 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 16:45:22 crc kubenswrapper[4744]: I1003 16:45:22.411948 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 16:45:22 crc kubenswrapper[4744]: I1003 16:45:22.463954 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 16:45:22 crc kubenswrapper[4744]: I1003 16:45:22.498912 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 16:45:22 crc kubenswrapper[4744]: I1003 16:45:22.906624 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f524b0a3-3650-4f2e-9a27-d74b82e3e06d" path="/var/lib/kubelet/pods/f524b0a3-3650-4f2e-9a27-d74b82e3e06d/volumes" Oct 03 16:45:23 crc kubenswrapper[4744]: I1003 16:45:23.168207 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"411dd208-5058-4f51-b5ab-4e4f8c77cd8e","Type":"ContainerStarted","Data":"6378fd9857476d1e4bcbf14d1a4d5b23149240e3339bb2861888afa16dcfe581"} Oct 03 16:45:23 crc kubenswrapper[4744]: I1003 16:45:23.168257 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 16:45:23 crc kubenswrapper[4744]: I1003 16:45:23.168270 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"411dd208-5058-4f51-b5ab-4e4f8c77cd8e","Type":"ContainerStarted","Data":"f93584cc754490792bd18f4cf0b75deb30a5568fbf615e0596a8169c7a450f28"} Oct 03 16:45:23 crc kubenswrapper[4744]: I1003 16:45:23.168422 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 16:45:23 crc kubenswrapper[4744]: I1003 16:45:23.608348 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:23 crc kubenswrapper[4744]: I1003 16:45:23.608708 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="proxy-httpd" containerID="cri-o://9407e0b7e8a5a7381cdcd46fca97ebd02c2143d6399cf5c16b4614d9dc90aedf" gracePeriod=30 Oct 03 16:45:23 crc kubenswrapper[4744]: I1003 16:45:23.608759 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="sg-core" containerID="cri-o://75122b72b34cab11d612bbe635e1de79825bf14c17fa5a4800fc8a9e02ca265c" gracePeriod=30 Oct 03 16:45:23 crc kubenswrapper[4744]: I1003 16:45:23.608762 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="ceilometer-notification-agent" containerID="cri-o://2be38fa7c5539b3eb33d6978b3f255b4fdaecebc80acb76b01c5f42261357222" gracePeriod=30 Oct 03 16:45:23 crc kubenswrapper[4744]: I1003 16:45:23.608708 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="ceilometer-central-agent" containerID="cri-o://ddd06e4de5ea54732fe7b9933e46b87deefcca81ee67200e32930825406411f5" gracePeriod=30 Oct 03 16:45:24 crc kubenswrapper[4744]: I1003 16:45:24.180825 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"411dd208-5058-4f51-b5ab-4e4f8c77cd8e","Type":"ContainerStarted","Data":"30313a0187b44c8f938ea876a5212d3535cea4910262941f2d25ef8e88eca0bb"} Oct 03 16:45:24 crc kubenswrapper[4744]: I1003 16:45:24.185280 4744 generic.go:334] "Generic (PLEG): container finished" podID="8fa06db6-8291-44c5-bd36-965c465ed970" containerID="9407e0b7e8a5a7381cdcd46fca97ebd02c2143d6399cf5c16b4614d9dc90aedf" exitCode=0 Oct 03 16:45:24 crc kubenswrapper[4744]: I1003 16:45:24.185342 4744 generic.go:334] "Generic (PLEG): container finished" podID="8fa06db6-8291-44c5-bd36-965c465ed970" containerID="75122b72b34cab11d612bbe635e1de79825bf14c17fa5a4800fc8a9e02ca265c" exitCode=2 Oct 03 16:45:24 crc kubenswrapper[4744]: I1003 16:45:24.185352 4744 generic.go:334] "Generic (PLEG): container finished" podID="8fa06db6-8291-44c5-bd36-965c465ed970" containerID="ddd06e4de5ea54732fe7b9933e46b87deefcca81ee67200e32930825406411f5" exitCode=0 Oct 03 16:45:24 crc kubenswrapper[4744]: I1003 16:45:24.185364 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fa06db6-8291-44c5-bd36-965c465ed970","Type":"ContainerDied","Data":"9407e0b7e8a5a7381cdcd46fca97ebd02c2143d6399cf5c16b4614d9dc90aedf"} Oct 03 16:45:24 crc kubenswrapper[4744]: I1003 16:45:24.185431 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fa06db6-8291-44c5-bd36-965c465ed970","Type":"ContainerDied","Data":"75122b72b34cab11d612bbe635e1de79825bf14c17fa5a4800fc8a9e02ca265c"} Oct 03 16:45:24 crc kubenswrapper[4744]: I1003 16:45:24.185441 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fa06db6-8291-44c5-bd36-965c465ed970","Type":"ContainerDied","Data":"ddd06e4de5ea54732fe7b9933e46b87deefcca81ee67200e32930825406411f5"} Oct 03 16:45:24 crc kubenswrapper[4744]: I1003 16:45:24.216611 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.216592837 podStartE2EDuration="3.216592837s" podCreationTimestamp="2025-10-03 16:45:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:45:24.207982738 +0000 UTC m=+1250.487858634" watchObservedRunningTime="2025-10-03 16:45:24.216592837 +0000 UTC m=+1250.496468733" Oct 03 16:45:25 crc kubenswrapper[4744]: I1003 16:45:25.200827 4744 generic.go:334] "Generic (PLEG): container finished" podID="8fa06db6-8291-44c5-bd36-965c465ed970" containerID="2be38fa7c5539b3eb33d6978b3f255b4fdaecebc80acb76b01c5f42261357222" exitCode=0 Oct 03 16:45:25 crc kubenswrapper[4744]: I1003 16:45:25.200945 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fa06db6-8291-44c5-bd36-965c465ed970","Type":"ContainerDied","Data":"2be38fa7c5539b3eb33d6978b3f255b4fdaecebc80acb76b01c5f42261357222"} Oct 03 16:45:25 crc kubenswrapper[4744]: I1003 16:45:25.201382 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 16:45:25 crc kubenswrapper[4744]: I1003 16:45:25.201398 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 16:45:25 crc kubenswrapper[4744]: I1003 16:45:25.303462 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 16:45:25 crc kubenswrapper[4744]: I1003 16:45:25.304738 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 16:45:28 crc kubenswrapper[4744]: I1003 16:45:28.469326 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.499237 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.614820 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-combined-ca-bundle\") pod \"8fa06db6-8291-44c5-bd36-965c465ed970\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.614958 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-scripts\") pod \"8fa06db6-8291-44c5-bd36-965c465ed970\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.614990 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-sg-core-conf-yaml\") pod \"8fa06db6-8291-44c5-bd36-965c465ed970\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.615022 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-config-data\") pod \"8fa06db6-8291-44c5-bd36-965c465ed970\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.615113 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fa06db6-8291-44c5-bd36-965c465ed970-run-httpd\") pod \"8fa06db6-8291-44c5-bd36-965c465ed970\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.615154 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j7qv\" (UniqueName: \"kubernetes.io/projected/8fa06db6-8291-44c5-bd36-965c465ed970-kube-api-access-8j7qv\") pod \"8fa06db6-8291-44c5-bd36-965c465ed970\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.615175 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fa06db6-8291-44c5-bd36-965c465ed970-log-httpd\") pod \"8fa06db6-8291-44c5-bd36-965c465ed970\" (UID: \"8fa06db6-8291-44c5-bd36-965c465ed970\") " Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.618389 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fa06db6-8291-44c5-bd36-965c465ed970-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8fa06db6-8291-44c5-bd36-965c465ed970" (UID: "8fa06db6-8291-44c5-bd36-965c465ed970"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.618877 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fa06db6-8291-44c5-bd36-965c465ed970-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8fa06db6-8291-44c5-bd36-965c465ed970" (UID: "8fa06db6-8291-44c5-bd36-965c465ed970"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.619389 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fa06db6-8291-44c5-bd36-965c465ed970-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.619416 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8fa06db6-8291-44c5-bd36-965c465ed970-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.620520 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-scripts" (OuterVolumeSpecName: "scripts") pod "8fa06db6-8291-44c5-bd36-965c465ed970" (UID: "8fa06db6-8291-44c5-bd36-965c465ed970"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.620889 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fa06db6-8291-44c5-bd36-965c465ed970-kube-api-access-8j7qv" (OuterVolumeSpecName: "kube-api-access-8j7qv") pod "8fa06db6-8291-44c5-bd36-965c465ed970" (UID: "8fa06db6-8291-44c5-bd36-965c465ed970"). InnerVolumeSpecName "kube-api-access-8j7qv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.668687 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8fa06db6-8291-44c5-bd36-965c465ed970" (UID: "8fa06db6-8291-44c5-bd36-965c465ed970"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.721837 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.721883 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.721902 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8j7qv\" (UniqueName: \"kubernetes.io/projected/8fa06db6-8291-44c5-bd36-965c465ed970-kube-api-access-8j7qv\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.732766 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-config-data" (OuterVolumeSpecName: "config-data") pod "8fa06db6-8291-44c5-bd36-965c465ed970" (UID: "8fa06db6-8291-44c5-bd36-965c465ed970"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.735104 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8fa06db6-8291-44c5-bd36-965c465ed970" (UID: "8fa06db6-8291-44c5-bd36-965c465ed970"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.827411 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.827481 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa06db6-8291-44c5-bd36-965c465ed970-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:30 crc kubenswrapper[4744]: I1003 16:45:30.880796 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.271466 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.272457 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8fa06db6-8291-44c5-bd36-965c465ed970","Type":"ContainerDied","Data":"eb978b80fd3637f6e97b676c232f1f70cffc536d8f6b0f46517f96f0acb640f9"} Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.272517 4744 scope.go:117] "RemoveContainer" containerID="9407e0b7e8a5a7381cdcd46fca97ebd02c2143d6399cf5c16b4614d9dc90aedf" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.273937 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zgrsm" event={"ID":"1cd09fa9-4967-49a1-a086-55405a0a681b","Type":"ContainerStarted","Data":"18c2a82b8e79a2029579d897d7f01f0583fdfbd27b3aa482d08a4d429783f1a8"} Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.299578 4744 scope.go:117] "RemoveContainer" containerID="75122b72b34cab11d612bbe635e1de79825bf14c17fa5a4800fc8a9e02ca265c" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.306613 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-zgrsm" podStartSLOduration=2.586089025 podStartE2EDuration="11.306595672s" podCreationTimestamp="2025-10-03 16:45:20 +0000 UTC" firstStartedPulling="2025-10-03 16:45:21.565815197 +0000 UTC m=+1247.845691093" lastFinishedPulling="2025-10-03 16:45:30.286321844 +0000 UTC m=+1256.566197740" observedRunningTime="2025-10-03 16:45:31.29824142 +0000 UTC m=+1257.578117326" watchObservedRunningTime="2025-10-03 16:45:31.306595672 +0000 UTC m=+1257.586471568" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.320384 4744 scope.go:117] "RemoveContainer" containerID="2be38fa7c5539b3eb33d6978b3f255b4fdaecebc80acb76b01c5f42261357222" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.321304 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.331524 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.342334 4744 scope.go:117] "RemoveContainer" containerID="ddd06e4de5ea54732fe7b9933e46b87deefcca81ee67200e32930825406411f5" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.352923 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:31 crc kubenswrapper[4744]: E1003 16:45:31.353378 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="proxy-httpd" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.353393 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="proxy-httpd" Oct 03 16:45:31 crc kubenswrapper[4744]: E1003 16:45:31.353420 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="ceilometer-central-agent" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.353428 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="ceilometer-central-agent" Oct 03 16:45:31 crc kubenswrapper[4744]: E1003 16:45:31.353441 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="sg-core" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.353447 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="sg-core" Oct 03 16:45:31 crc kubenswrapper[4744]: E1003 16:45:31.353470 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="ceilometer-notification-agent" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.353476 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="ceilometer-notification-agent" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.353667 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="sg-core" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.353682 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="ceilometer-notification-agent" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.353699 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="proxy-httpd" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.353708 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" containerName="ceilometer-central-agent" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.357663 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.361613 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.362662 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.367824 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.440445 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.440521 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq9gk\" (UniqueName: \"kubernetes.io/projected/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-kube-api-access-lq9gk\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.440579 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-log-httpd\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.440602 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-config-data\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.440635 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.440744 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-scripts\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.440764 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-run-httpd\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.517261 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.517332 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.544967 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.546101 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq9gk\" (UniqueName: \"kubernetes.io/projected/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-kube-api-access-lq9gk\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.546299 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-log-httpd\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.546359 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-config-data\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.546660 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-log-httpd\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.546854 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.547057 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-scripts\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.547103 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-run-httpd\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.547896 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-run-httpd\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.551299 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-config-data\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.551457 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.562665 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.563577 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-scripts\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.569454 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq9gk\" (UniqueName: \"kubernetes.io/projected/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-kube-api-access-lq9gk\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.570626 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " pod="openstack/ceilometer-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.581694 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 16:45:31 crc kubenswrapper[4744]: I1003 16:45:31.684615 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:45:32 crc kubenswrapper[4744]: I1003 16:45:32.185998 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:45:32 crc kubenswrapper[4744]: W1003 16:45:32.189477 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d74ef7e_3629_4819_a3d8_a8f7fd49b284.slice/crio-ec4ec5c6ad26b9832aaec6cab83972d68b3e643f27fd30ee94cf15644d4d0d0a WatchSource:0}: Error finding container ec4ec5c6ad26b9832aaec6cab83972d68b3e643f27fd30ee94cf15644d4d0d0a: Status 404 returned error can't find the container with id ec4ec5c6ad26b9832aaec6cab83972d68b3e643f27fd30ee94cf15644d4d0d0a Oct 03 16:45:32 crc kubenswrapper[4744]: I1003 16:45:32.282777 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d74ef7e-3629-4819-a3d8-a8f7fd49b284","Type":"ContainerStarted","Data":"ec4ec5c6ad26b9832aaec6cab83972d68b3e643f27fd30ee94cf15644d4d0d0a"} Oct 03 16:45:32 crc kubenswrapper[4744]: I1003 16:45:32.283220 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 16:45:32 crc kubenswrapper[4744]: I1003 16:45:32.283317 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 16:45:32 crc kubenswrapper[4744]: I1003 16:45:32.924916 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fa06db6-8291-44c5-bd36-965c465ed970" path="/var/lib/kubelet/pods/8fa06db6-8291-44c5-bd36-965c465ed970/volumes" Oct 03 16:45:33 crc kubenswrapper[4744]: I1003 16:45:33.292484 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d74ef7e-3629-4819-a3d8-a8f7fd49b284","Type":"ContainerStarted","Data":"7e1b1054724271bd934c3839f03d00046bc3cdc126777b683d26b07640b0c338"} Oct 03 16:45:34 crc kubenswrapper[4744]: I1003 16:45:34.223768 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 16:45:34 crc kubenswrapper[4744]: I1003 16:45:34.315994 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 16:45:34 crc kubenswrapper[4744]: I1003 16:45:34.322172 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d74ef7e-3629-4819-a3d8-a8f7fd49b284","Type":"ContainerStarted","Data":"503df61f5a4417b67b6bec262434d92ed9650d3845139b4ebbe0d4dcd3706c01"} Oct 03 16:45:35 crc kubenswrapper[4744]: I1003 16:45:35.334520 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d74ef7e-3629-4819-a3d8-a8f7fd49b284","Type":"ContainerStarted","Data":"58a8d3a5b2ee319eae5b88614209bd8af5fcdef4f06231b1770f1a2be95c566f"} Oct 03 16:45:37 crc kubenswrapper[4744]: I1003 16:45:37.361856 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d74ef7e-3629-4819-a3d8-a8f7fd49b284","Type":"ContainerStarted","Data":"34c8ada08e7d738040836f8abe9db2483c2abe4990e9c67d0804fb55756080ef"} Oct 03 16:45:37 crc kubenswrapper[4744]: I1003 16:45:37.362443 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 16:45:37 crc kubenswrapper[4744]: I1003 16:45:37.401839 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.444854982 podStartE2EDuration="6.401819688s" podCreationTimestamp="2025-10-03 16:45:31 +0000 UTC" firstStartedPulling="2025-10-03 16:45:32.192958155 +0000 UTC m=+1258.472834051" lastFinishedPulling="2025-10-03 16:45:36.149922861 +0000 UTC m=+1262.429798757" observedRunningTime="2025-10-03 16:45:37.390357646 +0000 UTC m=+1263.670233552" watchObservedRunningTime="2025-10-03 16:45:37.401819688 +0000 UTC m=+1263.681695594" Oct 03 16:45:40 crc kubenswrapper[4744]: I1003 16:45:40.001614 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 03 16:45:42 crc kubenswrapper[4744]: I1003 16:45:42.416213 4744 generic.go:334] "Generic (PLEG): container finished" podID="1cd09fa9-4967-49a1-a086-55405a0a681b" containerID="18c2a82b8e79a2029579d897d7f01f0583fdfbd27b3aa482d08a4d429783f1a8" exitCode=0 Oct 03 16:45:42 crc kubenswrapper[4744]: I1003 16:45:42.416328 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zgrsm" event={"ID":"1cd09fa9-4967-49a1-a086-55405a0a681b","Type":"ContainerDied","Data":"18c2a82b8e79a2029579d897d7f01f0583fdfbd27b3aa482d08a4d429783f1a8"} Oct 03 16:45:43 crc kubenswrapper[4744]: I1003 16:45:43.849877 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:43 crc kubenswrapper[4744]: I1003 16:45:43.924010 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-combined-ca-bundle\") pod \"1cd09fa9-4967-49a1-a086-55405a0a681b\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " Oct 03 16:45:43 crc kubenswrapper[4744]: I1003 16:45:43.924477 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsmvm\" (UniqueName: \"kubernetes.io/projected/1cd09fa9-4967-49a1-a086-55405a0a681b-kube-api-access-fsmvm\") pod \"1cd09fa9-4967-49a1-a086-55405a0a681b\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " Oct 03 16:45:43 crc kubenswrapper[4744]: I1003 16:45:43.924600 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-scripts\") pod \"1cd09fa9-4967-49a1-a086-55405a0a681b\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " Oct 03 16:45:43 crc kubenswrapper[4744]: I1003 16:45:43.924698 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-config-data\") pod \"1cd09fa9-4967-49a1-a086-55405a0a681b\" (UID: \"1cd09fa9-4967-49a1-a086-55405a0a681b\") " Oct 03 16:45:43 crc kubenswrapper[4744]: I1003 16:45:43.932657 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cd09fa9-4967-49a1-a086-55405a0a681b-kube-api-access-fsmvm" (OuterVolumeSpecName: "kube-api-access-fsmvm") pod "1cd09fa9-4967-49a1-a086-55405a0a681b" (UID: "1cd09fa9-4967-49a1-a086-55405a0a681b"). InnerVolumeSpecName "kube-api-access-fsmvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:45:43 crc kubenswrapper[4744]: I1003 16:45:43.934610 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-scripts" (OuterVolumeSpecName: "scripts") pod "1cd09fa9-4967-49a1-a086-55405a0a681b" (UID: "1cd09fa9-4967-49a1-a086-55405a0a681b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:43 crc kubenswrapper[4744]: I1003 16:45:43.972305 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-config-data" (OuterVolumeSpecName: "config-data") pod "1cd09fa9-4967-49a1-a086-55405a0a681b" (UID: "1cd09fa9-4967-49a1-a086-55405a0a681b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:43 crc kubenswrapper[4744]: I1003 16:45:43.976524 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1cd09fa9-4967-49a1-a086-55405a0a681b" (UID: "1cd09fa9-4967-49a1-a086-55405a0a681b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.030433 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsmvm\" (UniqueName: \"kubernetes.io/projected/1cd09fa9-4967-49a1-a086-55405a0a681b-kube-api-access-fsmvm\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.030530 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.030561 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.030585 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cd09fa9-4967-49a1-a086-55405a0a681b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.444182 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-zgrsm" event={"ID":"1cd09fa9-4967-49a1-a086-55405a0a681b","Type":"ContainerDied","Data":"645a07692dd7b8d15c29d16d28da0c05a077e77fafb0ebba682f1f84898ed1e2"} Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.444240 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="645a07692dd7b8d15c29d16d28da0c05a077e77fafb0ebba682f1f84898ed1e2" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.444263 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-zgrsm" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.563402 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 16:45:44 crc kubenswrapper[4744]: E1003 16:45:44.563899 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cd09fa9-4967-49a1-a086-55405a0a681b" containerName="nova-cell0-conductor-db-sync" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.563925 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cd09fa9-4967-49a1-a086-55405a0a681b" containerName="nova-cell0-conductor-db-sync" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.564151 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cd09fa9-4967-49a1-a086-55405a0a681b" containerName="nova-cell0-conductor-db-sync" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.564847 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.572221 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-p2fmp" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.572229 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.575168 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.748867 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da322563-8083-4e70-b339-a0c68c23e0c3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"da322563-8083-4e70-b339-a0c68c23e0c3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.748920 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da322563-8083-4e70-b339-a0c68c23e0c3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"da322563-8083-4e70-b339-a0c68c23e0c3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.749842 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcck8\" (UniqueName: \"kubernetes.io/projected/da322563-8083-4e70-b339-a0c68c23e0c3-kube-api-access-pcck8\") pod \"nova-cell0-conductor-0\" (UID: \"da322563-8083-4e70-b339-a0c68c23e0c3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.851297 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da322563-8083-4e70-b339-a0c68c23e0c3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"da322563-8083-4e70-b339-a0c68c23e0c3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.851354 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da322563-8083-4e70-b339-a0c68c23e0c3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"da322563-8083-4e70-b339-a0c68c23e0c3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.851566 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcck8\" (UniqueName: \"kubernetes.io/projected/da322563-8083-4e70-b339-a0c68c23e0c3-kube-api-access-pcck8\") pod \"nova-cell0-conductor-0\" (UID: \"da322563-8083-4e70-b339-a0c68c23e0c3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.857046 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da322563-8083-4e70-b339-a0c68c23e0c3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"da322563-8083-4e70-b339-a0c68c23e0c3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.857379 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da322563-8083-4e70-b339-a0c68c23e0c3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"da322563-8083-4e70-b339-a0c68c23e0c3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.876526 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcck8\" (UniqueName: \"kubernetes.io/projected/da322563-8083-4e70-b339-a0c68c23e0c3-kube-api-access-pcck8\") pod \"nova-cell0-conductor-0\" (UID: \"da322563-8083-4e70-b339-a0c68c23e0c3\") " pod="openstack/nova-cell0-conductor-0" Oct 03 16:45:44 crc kubenswrapper[4744]: I1003 16:45:44.889925 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 16:45:45 crc kubenswrapper[4744]: I1003 16:45:45.445778 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 16:45:45 crc kubenswrapper[4744]: I1003 16:45:45.461943 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"da322563-8083-4e70-b339-a0c68c23e0c3","Type":"ContainerStarted","Data":"d21ba0a6ae7f1869b3669e6fd72d79e07537df96a758efbfc414d1db63eb1b76"} Oct 03 16:45:46 crc kubenswrapper[4744]: I1003 16:45:46.477942 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"da322563-8083-4e70-b339-a0c68c23e0c3","Type":"ContainerStarted","Data":"286c216ad31a4a26990217d2cb041745927f6eadb7a79af05012960629310873"} Oct 03 16:45:46 crc kubenswrapper[4744]: I1003 16:45:46.478859 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 03 16:45:46 crc kubenswrapper[4744]: I1003 16:45:46.502075 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.502052227 podStartE2EDuration="2.502052227s" podCreationTimestamp="2025-10-03 16:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:45:46.499528253 +0000 UTC m=+1272.779404179" watchObservedRunningTime="2025-10-03 16:45:46.502052227 +0000 UTC m=+1272.781928163" Oct 03 16:45:54 crc kubenswrapper[4744]: I1003 16:45:54.941434 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.450828 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-g52js"] Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.452426 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.454824 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.459232 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.466341 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-g52js"] Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.520262 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-g52js\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.520343 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-scripts\") pod \"nova-cell0-cell-mapping-g52js\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.520391 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdrd7\" (UniqueName: \"kubernetes.io/projected/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-kube-api-access-cdrd7\") pod \"nova-cell0-cell-mapping-g52js\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.520443 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-config-data\") pod \"nova-cell0-cell-mapping-g52js\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.623732 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-scripts\") pod \"nova-cell0-cell-mapping-g52js\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.623789 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdrd7\" (UniqueName: \"kubernetes.io/projected/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-kube-api-access-cdrd7\") pod \"nova-cell0-cell-mapping-g52js\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.623846 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-config-data\") pod \"nova-cell0-cell-mapping-g52js\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.623927 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-g52js\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.633296 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-scripts\") pod \"nova-cell0-cell-mapping-g52js\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.641966 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-g52js\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.646297 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-config-data\") pod \"nova-cell0-cell-mapping-g52js\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.664259 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdrd7\" (UniqueName: \"kubernetes.io/projected/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-kube-api-access-cdrd7\") pod \"nova-cell0-cell-mapping-g52js\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.699607 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.701232 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.705777 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.716992 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.727350 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74cjs\" (UniqueName: \"kubernetes.io/projected/2373439b-a619-472d-9988-e01918981b4d-kube-api-access-74cjs\") pod \"nova-api-0\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " pod="openstack/nova-api-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.727413 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2373439b-a619-472d-9988-e01918981b4d-logs\") pod \"nova-api-0\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " pod="openstack/nova-api-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.727468 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2373439b-a619-472d-9988-e01918981b4d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " pod="openstack/nova-api-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.727577 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2373439b-a619-472d-9988-e01918981b4d-config-data\") pod \"nova-api-0\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " pod="openstack/nova-api-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.741473 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.743147 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.746737 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.785519 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.795983 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.797527 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.802448 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.844091 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.852724 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.899149 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgvc8\" (UniqueName: \"kubernetes.io/projected/133d6831-f643-4e6f-b245-ce4cca4a1b8c-kube-api-access-tgvc8\") pod \"nova-scheduler-0\" (UID: \"133d6831-f643-4e6f-b245-ce4cca4a1b8c\") " pod="openstack/nova-scheduler-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.899209 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2373439b-a619-472d-9988-e01918981b4d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " pod="openstack/nova-api-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.899271 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6x27\" (UniqueName: \"kubernetes.io/projected/ea42249b-a5a1-4728-8be6-8babf3c51aa2-kube-api-access-q6x27\") pod \"nova-metadata-0\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " pod="openstack/nova-metadata-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.899423 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2373439b-a619-472d-9988-e01918981b4d-config-data\") pod \"nova-api-0\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " pod="openstack/nova-api-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.899473 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea42249b-a5a1-4728-8be6-8babf3c51aa2-config-data\") pod \"nova-metadata-0\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " pod="openstack/nova-metadata-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.899527 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea42249b-a5a1-4728-8be6-8babf3c51aa2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " pod="openstack/nova-metadata-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.899551 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133d6831-f643-4e6f-b245-ce4cca4a1b8c-config-data\") pod \"nova-scheduler-0\" (UID: \"133d6831-f643-4e6f-b245-ce4cca4a1b8c\") " pod="openstack/nova-scheduler-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.899879 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea42249b-a5a1-4728-8be6-8babf3c51aa2-logs\") pod \"nova-metadata-0\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " pod="openstack/nova-metadata-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.899957 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74cjs\" (UniqueName: \"kubernetes.io/projected/2373439b-a619-472d-9988-e01918981b4d-kube-api-access-74cjs\") pod \"nova-api-0\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " pod="openstack/nova-api-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.900021 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2373439b-a619-472d-9988-e01918981b4d-logs\") pod \"nova-api-0\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " pod="openstack/nova-api-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.900051 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133d6831-f643-4e6f-b245-ce4cca4a1b8c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"133d6831-f643-4e6f-b245-ce4cca4a1b8c\") " pod="openstack/nova-scheduler-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.905810 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2373439b-a619-472d-9988-e01918981b4d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " pod="openstack/nova-api-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.921936 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2373439b-a619-472d-9988-e01918981b4d-logs\") pod \"nova-api-0\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " pod="openstack/nova-api-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.965233 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2373439b-a619-472d-9988-e01918981b4d-config-data\") pod \"nova-api-0\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " pod="openstack/nova-api-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.967804 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.969113 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.972858 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.974871 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74cjs\" (UniqueName: \"kubernetes.io/projected/2373439b-a619-472d-9988-e01918981b4d-kube-api-access-74cjs\") pod \"nova-api-0\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " pod="openstack/nova-api-0" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.978707 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs"] Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.980246 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:55 crc kubenswrapper[4744]: I1003 16:45:55.992314 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.000760 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs"] Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.005779 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea42249b-a5a1-4728-8be6-8babf3c51aa2-logs\") pod \"nova-metadata-0\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " pod="openstack/nova-metadata-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.005878 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133d6831-f643-4e6f-b245-ce4cca4a1b8c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"133d6831-f643-4e6f-b245-ce4cca4a1b8c\") " pod="openstack/nova-scheduler-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.005905 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgvc8\" (UniqueName: \"kubernetes.io/projected/133d6831-f643-4e6f-b245-ce4cca4a1b8c-kube-api-access-tgvc8\") pod \"nova-scheduler-0\" (UID: \"133d6831-f643-4e6f-b245-ce4cca4a1b8c\") " pod="openstack/nova-scheduler-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.005939 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6x27\" (UniqueName: \"kubernetes.io/projected/ea42249b-a5a1-4728-8be6-8babf3c51aa2-kube-api-access-q6x27\") pod \"nova-metadata-0\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " pod="openstack/nova-metadata-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.006010 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea42249b-a5a1-4728-8be6-8babf3c51aa2-config-data\") pod \"nova-metadata-0\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " pod="openstack/nova-metadata-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.006036 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea42249b-a5a1-4728-8be6-8babf3c51aa2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " pod="openstack/nova-metadata-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.006053 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133d6831-f643-4e6f-b245-ce4cca4a1b8c-config-data\") pod \"nova-scheduler-0\" (UID: \"133d6831-f643-4e6f-b245-ce4cca4a1b8c\") " pod="openstack/nova-scheduler-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.010361 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea42249b-a5a1-4728-8be6-8babf3c51aa2-logs\") pod \"nova-metadata-0\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " pod="openstack/nova-metadata-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.022746 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133d6831-f643-4e6f-b245-ce4cca4a1b8c-config-data\") pod \"nova-scheduler-0\" (UID: \"133d6831-f643-4e6f-b245-ce4cca4a1b8c\") " pod="openstack/nova-scheduler-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.023047 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea42249b-a5a1-4728-8be6-8babf3c51aa2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " pod="openstack/nova-metadata-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.024933 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133d6831-f643-4e6f-b245-ce4cca4a1b8c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"133d6831-f643-4e6f-b245-ce4cca4a1b8c\") " pod="openstack/nova-scheduler-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.025620 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea42249b-a5a1-4728-8be6-8babf3c51aa2-config-data\") pod \"nova-metadata-0\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " pod="openstack/nova-metadata-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.026038 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6x27\" (UniqueName: \"kubernetes.io/projected/ea42249b-a5a1-4728-8be6-8babf3c51aa2-kube-api-access-q6x27\") pod \"nova-metadata-0\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " pod="openstack/nova-metadata-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.026763 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgvc8\" (UniqueName: \"kubernetes.io/projected/133d6831-f643-4e6f-b245-ce4cca4a1b8c-kube-api-access-tgvc8\") pod \"nova-scheduler-0\" (UID: \"133d6831-f643-4e6f-b245-ce4cca4a1b8c\") " pod="openstack/nova-scheduler-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.081813 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.096582 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.106570 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.108611 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-config\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.108652 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2528f1c7-4a26-472c-84a5-c5d0a90c5ece\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.108704 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkw7d\" (UniqueName: \"kubernetes.io/projected/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-kube-api-access-zkw7d\") pod \"nova-cell1-novncproxy-0\" (UID: \"2528f1c7-4a26-472c-84a5-c5d0a90c5ece\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.108746 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq4lz\" (UniqueName: \"kubernetes.io/projected/29c874fa-abab-46d4-99e5-b44e6332f71d-kube-api-access-mq4lz\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.108783 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-ovsdbserver-nb\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.108810 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-ovsdbserver-sb\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.109365 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-dns-swift-storage-0\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.109485 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2528f1c7-4a26-472c-84a5-c5d0a90c5ece\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.109532 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-dns-svc\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.211543 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkw7d\" (UniqueName: \"kubernetes.io/projected/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-kube-api-access-zkw7d\") pod \"nova-cell1-novncproxy-0\" (UID: \"2528f1c7-4a26-472c-84a5-c5d0a90c5ece\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.211904 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq4lz\" (UniqueName: \"kubernetes.io/projected/29c874fa-abab-46d4-99e5-b44e6332f71d-kube-api-access-mq4lz\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.212226 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-ovsdbserver-nb\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.212252 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-ovsdbserver-sb\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.212364 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-dns-swift-storage-0\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.212393 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2528f1c7-4a26-472c-84a5-c5d0a90c5ece\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.212418 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-dns-svc\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.212449 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-config\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.212481 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2528f1c7-4a26-472c-84a5-c5d0a90c5ece\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.213686 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-dns-swift-storage-0\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.213765 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-ovsdbserver-nb\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.213906 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-ovsdbserver-sb\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.213957 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-dns-svc\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.216574 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2528f1c7-4a26-472c-84a5-c5d0a90c5ece\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.217143 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-config\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.219074 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2528f1c7-4a26-472c-84a5-c5d0a90c5ece\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.227041 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkw7d\" (UniqueName: \"kubernetes.io/projected/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-kube-api-access-zkw7d\") pod \"nova-cell1-novncproxy-0\" (UID: \"2528f1c7-4a26-472c-84a5-c5d0a90c5ece\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.238378 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq4lz\" (UniqueName: \"kubernetes.io/projected/29c874fa-abab-46d4-99e5-b44e6332f71d-kube-api-access-mq4lz\") pod \"dnsmasq-dns-7d5fbbb8c5-xjwgs\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.418624 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.426289 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.433555 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-g52js"] Oct 03 16:45:56 crc kubenswrapper[4744]: W1003 16:45:56.441740 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4faff3af_3870_4a9b_9fcf_91f0ecc30a8b.slice/crio-3000ac1f8fee821a1211c464076e92ae28c6028865c329309219800f384b4440 WatchSource:0}: Error finding container 3000ac1f8fee821a1211c464076e92ae28c6028865c329309219800f384b4440: Status 404 returned error can't find the container with id 3000ac1f8fee821a1211c464076e92ae28c6028865c329309219800f384b4440 Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.656612 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g52js" event={"ID":"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b","Type":"ContainerStarted","Data":"3000ac1f8fee821a1211c464076e92ae28c6028865c329309219800f384b4440"} Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.665844 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qhgd6"] Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.667331 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.673671 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.673754 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.694956 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qhgd6"] Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.725912 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-scripts\") pod \"nova-cell1-conductor-db-sync-qhgd6\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.725962 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-qhgd6\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.725992 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6bmw\" (UniqueName: \"kubernetes.io/projected/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-kube-api-access-z6bmw\") pod \"nova-cell1-conductor-db-sync-qhgd6\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.726119 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-config-data\") pod \"nova-cell1-conductor-db-sync-qhgd6\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:45:56 crc kubenswrapper[4744]: W1003 16:45:56.730389 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea42249b_a5a1_4728_8be6_8babf3c51aa2.slice/crio-e0f923a0cc6165a2f33f087670219971545518476e8d2bc29526480aec44f3d2 WatchSource:0}: Error finding container e0f923a0cc6165a2f33f087670219971545518476e8d2bc29526480aec44f3d2: Status 404 returned error can't find the container with id e0f923a0cc6165a2f33f087670219971545518476e8d2bc29526480aec44f3d2 Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.738563 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.770939 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.828964 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-config-data\") pod \"nova-cell1-conductor-db-sync-qhgd6\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.829042 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-scripts\") pod \"nova-cell1-conductor-db-sync-qhgd6\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.829064 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-qhgd6\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.829091 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6bmw\" (UniqueName: \"kubernetes.io/projected/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-kube-api-access-z6bmw\") pod \"nova-cell1-conductor-db-sync-qhgd6\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.834847 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-config-data\") pod \"nova-cell1-conductor-db-sync-qhgd6\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.837988 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-qhgd6\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.838855 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-scripts\") pod \"nova-cell1-conductor-db-sync-qhgd6\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.839280 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.846802 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6bmw\" (UniqueName: \"kubernetes.io/projected/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-kube-api-access-z6bmw\") pod \"nova-cell1-conductor-db-sync-qhgd6\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:45:56 crc kubenswrapper[4744]: I1003 16:45:56.936218 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:45:57 crc kubenswrapper[4744]: I1003 16:45:57.006514 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 16:45:57 crc kubenswrapper[4744]: I1003 16:45:57.013301 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs"] Oct 03 16:45:57 crc kubenswrapper[4744]: W1003 16:45:57.041799 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2528f1c7_4a26_472c_84a5_c5d0a90c5ece.slice/crio-040941ba0767c5e086195037328eb12e86a407d56256cef58adeb58326778b27 WatchSource:0}: Error finding container 040941ba0767c5e086195037328eb12e86a407d56256cef58adeb58326778b27: Status 404 returned error can't find the container with id 040941ba0767c5e086195037328eb12e86a407d56256cef58adeb58326778b27 Oct 03 16:45:57 crc kubenswrapper[4744]: I1003 16:45:57.454434 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qhgd6"] Oct 03 16:45:57 crc kubenswrapper[4744]: W1003 16:45:57.480545 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb1cb635_71ed_4d6f_89ba_b2dd9342a55a.slice/crio-df17f1e676ef4b7965a95b6a53a3f7781a64362021ff4bf4973205008cf20db8 WatchSource:0}: Error finding container df17f1e676ef4b7965a95b6a53a3f7781a64362021ff4bf4973205008cf20db8: Status 404 returned error can't find the container with id df17f1e676ef4b7965a95b6a53a3f7781a64362021ff4bf4973205008cf20db8 Oct 03 16:45:57 crc kubenswrapper[4744]: I1003 16:45:57.677328 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qhgd6" event={"ID":"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a","Type":"ContainerStarted","Data":"df17f1e676ef4b7965a95b6a53a3f7781a64362021ff4bf4973205008cf20db8"} Oct 03 16:45:57 crc kubenswrapper[4744]: I1003 16:45:57.679468 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ea42249b-a5a1-4728-8be6-8babf3c51aa2","Type":"ContainerStarted","Data":"e0f923a0cc6165a2f33f087670219971545518476e8d2bc29526480aec44f3d2"} Oct 03 16:45:57 crc kubenswrapper[4744]: I1003 16:45:57.682928 4744 generic.go:334] "Generic (PLEG): container finished" podID="29c874fa-abab-46d4-99e5-b44e6332f71d" containerID="54759bc2ac35a0adf4ad31b0b2b040444896aa440fd9a749dc4119f740569c27" exitCode=0 Oct 03 16:45:57 crc kubenswrapper[4744]: I1003 16:45:57.683010 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" event={"ID":"29c874fa-abab-46d4-99e5-b44e6332f71d","Type":"ContainerDied","Data":"54759bc2ac35a0adf4ad31b0b2b040444896aa440fd9a749dc4119f740569c27"} Oct 03 16:45:57 crc kubenswrapper[4744]: I1003 16:45:57.683040 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" event={"ID":"29c874fa-abab-46d4-99e5-b44e6332f71d","Type":"ContainerStarted","Data":"a23ffcd21f98b404999bfdd5a18c5b44cff652def773b7402b44461350d1bee8"} Oct 03 16:45:57 crc kubenswrapper[4744]: I1003 16:45:57.690050 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2373439b-a619-472d-9988-e01918981b4d","Type":"ContainerStarted","Data":"fc8abf325495881da3205a82aee0ba07448743113a2290b280ac9a932fe8b102"} Oct 03 16:45:57 crc kubenswrapper[4744]: I1003 16:45:57.693270 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g52js" event={"ID":"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b","Type":"ContainerStarted","Data":"f51e3986742213bb93feed75ec3949a025047f40d33b489bdc97accde4460cab"} Oct 03 16:45:57 crc kubenswrapper[4744]: I1003 16:45:57.694642 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"133d6831-f643-4e6f-b245-ce4cca4a1b8c","Type":"ContainerStarted","Data":"3dc654619212b2f17f51aab006ee84edc790af1d32f5406cd4de5d5594355f79"} Oct 03 16:45:57 crc kubenswrapper[4744]: I1003 16:45:57.695711 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2528f1c7-4a26-472c-84a5-c5d0a90c5ece","Type":"ContainerStarted","Data":"040941ba0767c5e086195037328eb12e86a407d56256cef58adeb58326778b27"} Oct 03 16:45:57 crc kubenswrapper[4744]: I1003 16:45:57.775216 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-g52js" podStartSLOduration=2.7751962470000002 podStartE2EDuration="2.775196247s" podCreationTimestamp="2025-10-03 16:45:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:45:57.735702463 +0000 UTC m=+1284.015578359" watchObservedRunningTime="2025-10-03 16:45:57.775196247 +0000 UTC m=+1284.055072133" Oct 03 16:45:58 crc kubenswrapper[4744]: I1003 16:45:58.713394 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qhgd6" event={"ID":"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a","Type":"ContainerStarted","Data":"20a31bbd941873591b0299af896e8dd8cc01fae49345187d13a006528fb702a5"} Oct 03 16:45:58 crc kubenswrapper[4744]: I1003 16:45:58.717756 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" event={"ID":"29c874fa-abab-46d4-99e5-b44e6332f71d","Type":"ContainerStarted","Data":"c84611e4217cd7d9a6a73bfcb2bfef411bb121c02dc022732dd5af8dbc5eef5d"} Oct 03 16:45:58 crc kubenswrapper[4744]: I1003 16:45:58.717963 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:45:58 crc kubenswrapper[4744]: I1003 16:45:58.731822 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-qhgd6" podStartSLOduration=2.731794406 podStartE2EDuration="2.731794406s" podCreationTimestamp="2025-10-03 16:45:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:45:58.727740063 +0000 UTC m=+1285.007615959" watchObservedRunningTime="2025-10-03 16:45:58.731794406 +0000 UTC m=+1285.011670342" Oct 03 16:45:58 crc kubenswrapper[4744]: I1003 16:45:58.753646 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" podStartSLOduration=3.7536244610000002 podStartE2EDuration="3.753624461s" podCreationTimestamp="2025-10-03 16:45:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:45:58.74414384 +0000 UTC m=+1285.024019736" watchObservedRunningTime="2025-10-03 16:45:58.753624461 +0000 UTC m=+1285.033500377" Oct 03 16:45:59 crc kubenswrapper[4744]: I1003 16:45:59.162321 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:45:59 crc kubenswrapper[4744]: I1003 16:45:59.180081 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 16:46:00 crc kubenswrapper[4744]: I1003 16:46:00.745196 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2373439b-a619-472d-9988-e01918981b4d","Type":"ContainerStarted","Data":"e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e"} Oct 03 16:46:00 crc kubenswrapper[4744]: I1003 16:46:00.745902 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2373439b-a619-472d-9988-e01918981b4d","Type":"ContainerStarted","Data":"351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713"} Oct 03 16:46:00 crc kubenswrapper[4744]: I1003 16:46:00.746942 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"133d6831-f643-4e6f-b245-ce4cca4a1b8c","Type":"ContainerStarted","Data":"922b19d530b25702f80d628149b1977891e6a4399d41d708e3cfd9744f7dfa22"} Oct 03 16:46:00 crc kubenswrapper[4744]: I1003 16:46:00.749178 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2528f1c7-4a26-472c-84a5-c5d0a90c5ece","Type":"ContainerStarted","Data":"d6ac5af47e78703484a6564224485d108b26016a512fe43491200033e1079231"} Oct 03 16:46:00 crc kubenswrapper[4744]: I1003 16:46:00.749342 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="2528f1c7-4a26-472c-84a5-c5d0a90c5ece" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://d6ac5af47e78703484a6564224485d108b26016a512fe43491200033e1079231" gracePeriod=30 Oct 03 16:46:00 crc kubenswrapper[4744]: I1003 16:46:00.751965 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ea42249b-a5a1-4728-8be6-8babf3c51aa2","Type":"ContainerStarted","Data":"2d3ec7a1b23df0aeefde51845f97da1a41107b85563c69c240d72fdd219d0625"} Oct 03 16:46:00 crc kubenswrapper[4744]: I1003 16:46:00.752005 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ea42249b-a5a1-4728-8be6-8babf3c51aa2","Type":"ContainerStarted","Data":"bd91db54296bad89568c91d52e226066828dcab4426f18c5920f65119bcf722f"} Oct 03 16:46:00 crc kubenswrapper[4744]: I1003 16:46:00.752124 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ea42249b-a5a1-4728-8be6-8babf3c51aa2" containerName="nova-metadata-log" containerID="cri-o://bd91db54296bad89568c91d52e226066828dcab4426f18c5920f65119bcf722f" gracePeriod=30 Oct 03 16:46:00 crc kubenswrapper[4744]: I1003 16:46:00.752255 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ea42249b-a5a1-4728-8be6-8babf3c51aa2" containerName="nova-metadata-metadata" containerID="cri-o://2d3ec7a1b23df0aeefde51845f97da1a41107b85563c69c240d72fdd219d0625" gracePeriod=30 Oct 03 16:46:00 crc kubenswrapper[4744]: I1003 16:46:00.778363 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.534948489 podStartE2EDuration="5.778342514s" podCreationTimestamp="2025-10-03 16:45:55 +0000 UTC" firstStartedPulling="2025-10-03 16:45:56.846357194 +0000 UTC m=+1283.126233100" lastFinishedPulling="2025-10-03 16:46:00.089751229 +0000 UTC m=+1286.369627125" observedRunningTime="2025-10-03 16:46:00.766872022 +0000 UTC m=+1287.046747918" watchObservedRunningTime="2025-10-03 16:46:00.778342514 +0000 UTC m=+1287.058218430" Oct 03 16:46:00 crc kubenswrapper[4744]: I1003 16:46:00.784990 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.7435102909999998 podStartE2EDuration="5.784974513s" podCreationTimestamp="2025-10-03 16:45:55 +0000 UTC" firstStartedPulling="2025-10-03 16:45:57.047721863 +0000 UTC m=+1283.327597749" lastFinishedPulling="2025-10-03 16:46:00.089186045 +0000 UTC m=+1286.369061971" observedRunningTime="2025-10-03 16:46:00.784254364 +0000 UTC m=+1287.064130270" watchObservedRunningTime="2025-10-03 16:46:00.784974513 +0000 UTC m=+1287.064850409" Oct 03 16:46:00 crc kubenswrapper[4744]: I1003 16:46:00.809141 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.426280928 podStartE2EDuration="5.809118837s" podCreationTimestamp="2025-10-03 16:45:55 +0000 UTC" firstStartedPulling="2025-10-03 16:45:56.734133002 +0000 UTC m=+1283.014008898" lastFinishedPulling="2025-10-03 16:46:00.116970911 +0000 UTC m=+1286.396846807" observedRunningTime="2025-10-03 16:46:00.801259747 +0000 UTC m=+1287.081135643" watchObservedRunningTime="2025-10-03 16:46:00.809118837 +0000 UTC m=+1287.088994733" Oct 03 16:46:00 crc kubenswrapper[4744]: I1003 16:46:00.825071 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.5087508830000003 podStartE2EDuration="5.825051822s" podCreationTimestamp="2025-10-03 16:45:55 +0000 UTC" firstStartedPulling="2025-10-03 16:45:56.772854885 +0000 UTC m=+1283.052730781" lastFinishedPulling="2025-10-03 16:46:00.089155794 +0000 UTC m=+1286.369031720" observedRunningTime="2025-10-03 16:46:00.81517412 +0000 UTC m=+1287.095050016" watchObservedRunningTime="2025-10-03 16:46:00.825051822 +0000 UTC m=+1287.104927718" Oct 03 16:46:01 crc kubenswrapper[4744]: I1003 16:46:01.097820 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 16:46:01 crc kubenswrapper[4744]: I1003 16:46:01.098195 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 16:46:01 crc kubenswrapper[4744]: I1003 16:46:01.107400 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 03 16:46:01 crc kubenswrapper[4744]: I1003 16:46:01.419725 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:01 crc kubenswrapper[4744]: I1003 16:46:01.704400 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 03 16:46:01 crc kubenswrapper[4744]: I1003 16:46:01.765061 4744 generic.go:334] "Generic (PLEG): container finished" podID="ea42249b-a5a1-4728-8be6-8babf3c51aa2" containerID="bd91db54296bad89568c91d52e226066828dcab4426f18c5920f65119bcf722f" exitCode=143 Oct 03 16:46:01 crc kubenswrapper[4744]: I1003 16:46:01.766253 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ea42249b-a5a1-4728-8be6-8babf3c51aa2","Type":"ContainerDied","Data":"bd91db54296bad89568c91d52e226066828dcab4426f18c5920f65119bcf722f"} Oct 03 16:46:03 crc kubenswrapper[4744]: I1003 16:46:03.787969 4744 generic.go:334] "Generic (PLEG): container finished" podID="4faff3af-3870-4a9b-9fcf-91f0ecc30a8b" containerID="f51e3986742213bb93feed75ec3949a025047f40d33b489bdc97accde4460cab" exitCode=0 Oct 03 16:46:03 crc kubenswrapper[4744]: I1003 16:46:03.788065 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g52js" event={"ID":"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b","Type":"ContainerDied","Data":"f51e3986742213bb93feed75ec3949a025047f40d33b489bdc97accde4460cab"} Oct 03 16:46:04 crc kubenswrapper[4744]: I1003 16:46:04.802023 4744 generic.go:334] "Generic (PLEG): container finished" podID="bb1cb635-71ed-4d6f-89ba-b2dd9342a55a" containerID="20a31bbd941873591b0299af896e8dd8cc01fae49345187d13a006528fb702a5" exitCode=0 Oct 03 16:46:04 crc kubenswrapper[4744]: I1003 16:46:04.802100 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qhgd6" event={"ID":"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a","Type":"ContainerDied","Data":"20a31bbd941873591b0299af896e8dd8cc01fae49345187d13a006528fb702a5"} Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.248164 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.258170 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.258397 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="e104e42e-cc57-4233-b7a5-2755705d18f7" containerName="kube-state-metrics" containerID="cri-o://3090f3f9894d8ef9f8d9519dbde832d68952a7ea5dc64f2385cdfa0d661cd64e" gracePeriod=30 Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.332898 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdrd7\" (UniqueName: \"kubernetes.io/projected/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-kube-api-access-cdrd7\") pod \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.333103 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-combined-ca-bundle\") pod \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.333160 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-config-data\") pod \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.333252 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-scripts\") pod \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\" (UID: \"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b\") " Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.340713 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-scripts" (OuterVolumeSpecName: "scripts") pod "4faff3af-3870-4a9b-9fcf-91f0ecc30a8b" (UID: "4faff3af-3870-4a9b-9fcf-91f0ecc30a8b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.346682 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-kube-api-access-cdrd7" (OuterVolumeSpecName: "kube-api-access-cdrd7") pod "4faff3af-3870-4a9b-9fcf-91f0ecc30a8b" (UID: "4faff3af-3870-4a9b-9fcf-91f0ecc30a8b"). InnerVolumeSpecName "kube-api-access-cdrd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.366858 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4faff3af-3870-4a9b-9fcf-91f0ecc30a8b" (UID: "4faff3af-3870-4a9b-9fcf-91f0ecc30a8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.369350 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-config-data" (OuterVolumeSpecName: "config-data") pod "4faff3af-3870-4a9b-9fcf-91f0ecc30a8b" (UID: "4faff3af-3870-4a9b-9fcf-91f0ecc30a8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.435733 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.435770 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.435780 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdrd7\" (UniqueName: \"kubernetes.io/projected/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-kube-api-access-cdrd7\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.435791 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.713254 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.813992 4744 generic.go:334] "Generic (PLEG): container finished" podID="e104e42e-cc57-4233-b7a5-2755705d18f7" containerID="3090f3f9894d8ef9f8d9519dbde832d68952a7ea5dc64f2385cdfa0d661cd64e" exitCode=2 Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.814197 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e104e42e-cc57-4233-b7a5-2755705d18f7","Type":"ContainerDied","Data":"3090f3f9894d8ef9f8d9519dbde832d68952a7ea5dc64f2385cdfa0d661cd64e"} Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.814454 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e104e42e-cc57-4233-b7a5-2755705d18f7","Type":"ContainerDied","Data":"983936e1ca08042b3c479d3bd4e40443aa75778f55c0eba77cb4e5ccdaca0de0"} Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.814273 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.814482 4744 scope.go:117] "RemoveContainer" containerID="3090f3f9894d8ef9f8d9519dbde832d68952a7ea5dc64f2385cdfa0d661cd64e" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.818535 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g52js" event={"ID":"4faff3af-3870-4a9b-9fcf-91f0ecc30a8b","Type":"ContainerDied","Data":"3000ac1f8fee821a1211c464076e92ae28c6028865c329309219800f384b4440"} Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.818572 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3000ac1f8fee821a1211c464076e92ae28c6028865c329309219800f384b4440" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.818604 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g52js" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.844252 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cr2f7\" (UniqueName: \"kubernetes.io/projected/e104e42e-cc57-4233-b7a5-2755705d18f7-kube-api-access-cr2f7\") pod \"e104e42e-cc57-4233-b7a5-2755705d18f7\" (UID: \"e104e42e-cc57-4233-b7a5-2755705d18f7\") " Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.850934 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e104e42e-cc57-4233-b7a5-2755705d18f7-kube-api-access-cr2f7" (OuterVolumeSpecName: "kube-api-access-cr2f7") pod "e104e42e-cc57-4233-b7a5-2755705d18f7" (UID: "e104e42e-cc57-4233-b7a5-2755705d18f7"). InnerVolumeSpecName "kube-api-access-cr2f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.852412 4744 scope.go:117] "RemoveContainer" containerID="3090f3f9894d8ef9f8d9519dbde832d68952a7ea5dc64f2385cdfa0d661cd64e" Oct 03 16:46:05 crc kubenswrapper[4744]: E1003 16:46:05.853383 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3090f3f9894d8ef9f8d9519dbde832d68952a7ea5dc64f2385cdfa0d661cd64e\": container with ID starting with 3090f3f9894d8ef9f8d9519dbde832d68952a7ea5dc64f2385cdfa0d661cd64e not found: ID does not exist" containerID="3090f3f9894d8ef9f8d9519dbde832d68952a7ea5dc64f2385cdfa0d661cd64e" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.853431 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3090f3f9894d8ef9f8d9519dbde832d68952a7ea5dc64f2385cdfa0d661cd64e"} err="failed to get container status \"3090f3f9894d8ef9f8d9519dbde832d68952a7ea5dc64f2385cdfa0d661cd64e\": rpc error: code = NotFound desc = could not find container \"3090f3f9894d8ef9f8d9519dbde832d68952a7ea5dc64f2385cdfa0d661cd64e\": container with ID starting with 3090f3f9894d8ef9f8d9519dbde832d68952a7ea5dc64f2385cdfa0d661cd64e not found: ID does not exist" Oct 03 16:46:05 crc kubenswrapper[4744]: I1003 16:46:05.947583 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cr2f7\" (UniqueName: \"kubernetes.io/projected/e104e42e-cc57-4233-b7a5-2755705d18f7-kube-api-access-cr2f7\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.029578 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.029794 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2373439b-a619-472d-9988-e01918981b4d" containerName="nova-api-log" containerID="cri-o://351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713" gracePeriod=30 Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.030123 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2373439b-a619-472d-9988-e01918981b4d" containerName="nova-api-api" containerID="cri-o://e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e" gracePeriod=30 Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.045012 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.045215 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="133d6831-f643-4e6f-b245-ce4cca4a1b8c" containerName="nova-scheduler-scheduler" containerID="cri-o://922b19d530b25702f80d628149b1977891e6a4399d41d708e3cfd9744f7dfa22" gracePeriod=30 Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.364089 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.379281 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.393674 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.412681 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 16:46:06 crc kubenswrapper[4744]: E1003 16:46:06.413135 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e104e42e-cc57-4233-b7a5-2755705d18f7" containerName="kube-state-metrics" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.413158 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e104e42e-cc57-4233-b7a5-2755705d18f7" containerName="kube-state-metrics" Oct 03 16:46:06 crc kubenswrapper[4744]: E1003 16:46:06.413183 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb1cb635-71ed-4d6f-89ba-b2dd9342a55a" containerName="nova-cell1-conductor-db-sync" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.413190 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb1cb635-71ed-4d6f-89ba-b2dd9342a55a" containerName="nova-cell1-conductor-db-sync" Oct 03 16:46:06 crc kubenswrapper[4744]: E1003 16:46:06.413214 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4faff3af-3870-4a9b-9fcf-91f0ecc30a8b" containerName="nova-manage" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.413221 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4faff3af-3870-4a9b-9fcf-91f0ecc30a8b" containerName="nova-manage" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.413390 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e104e42e-cc57-4233-b7a5-2755705d18f7" containerName="kube-state-metrics" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.413411 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4faff3af-3870-4a9b-9fcf-91f0ecc30a8b" containerName="nova-manage" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.413425 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb1cb635-71ed-4d6f-89ba-b2dd9342a55a" containerName="nova-cell1-conductor-db-sync" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.414208 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.416631 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.416655 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.429099 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.443275 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.464233 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6bmw\" (UniqueName: \"kubernetes.io/projected/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-kube-api-access-z6bmw\") pod \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.464341 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-scripts\") pod \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.464431 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-combined-ca-bundle\") pod \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.464485 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-config-data\") pod \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\" (UID: \"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a\") " Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.464922 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfld7\" (UniqueName: \"kubernetes.io/projected/6c69a90c-bccf-4209-94e1-fe80d4914377-kube-api-access-pfld7\") pod \"kube-state-metrics-0\" (UID: \"6c69a90c-bccf-4209-94e1-fe80d4914377\") " pod="openstack/kube-state-metrics-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.465054 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6c69a90c-bccf-4209-94e1-fe80d4914377-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6c69a90c-bccf-4209-94e1-fe80d4914377\") " pod="openstack/kube-state-metrics-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.465103 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c69a90c-bccf-4209-94e1-fe80d4914377-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6c69a90c-bccf-4209-94e1-fe80d4914377\") " pod="openstack/kube-state-metrics-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.465152 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c69a90c-bccf-4209-94e1-fe80d4914377-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6c69a90c-bccf-4209-94e1-fe80d4914377\") " pod="openstack/kube-state-metrics-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.499689 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-scripts" (OuterVolumeSpecName: "scripts") pod "bb1cb635-71ed-4d6f-89ba-b2dd9342a55a" (UID: "bb1cb635-71ed-4d6f-89ba-b2dd9342a55a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:06 crc kubenswrapper[4744]: E1003 16:46:06.500419 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2373439b_a619_472d_9988_e01918981b4d.slice/crio-e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode104e42e_cc57_4233_b7a5_2755705d18f7.slice/crio-983936e1ca08042b3c479d3bd4e40443aa75778f55c0eba77cb4e5ccdaca0de0\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2373439b_a619_472d_9988_e01918981b4d.slice/crio-conmon-e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode104e42e_cc57_4233_b7a5_2755705d18f7.slice\": RecentStats: unable to find data in memory cache]" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.502348 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb1cb635-71ed-4d6f-89ba-b2dd9342a55a" (UID: "bb1cb635-71ed-4d6f-89ba-b2dd9342a55a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.511094 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5865f9d689-cwtl8"] Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.511314 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" podUID="686a87c3-18f7-4578-a07c-ff9fb5a766cd" containerName="dnsmasq-dns" containerID="cri-o://53872c77e10672dc21a6ea541bf521cc51f9325fd6745aca8761d4721a9c4097" gracePeriod=10 Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.534722 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-config-data" (OuterVolumeSpecName: "config-data") pod "bb1cb635-71ed-4d6f-89ba-b2dd9342a55a" (UID: "bb1cb635-71ed-4d6f-89ba-b2dd9342a55a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.571629 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c69a90c-bccf-4209-94e1-fe80d4914377-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6c69a90c-bccf-4209-94e1-fe80d4914377\") " pod="openstack/kube-state-metrics-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.571703 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c69a90c-bccf-4209-94e1-fe80d4914377-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6c69a90c-bccf-4209-94e1-fe80d4914377\") " pod="openstack/kube-state-metrics-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.571801 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfld7\" (UniqueName: \"kubernetes.io/projected/6c69a90c-bccf-4209-94e1-fe80d4914377-kube-api-access-pfld7\") pod \"kube-state-metrics-0\" (UID: \"6c69a90c-bccf-4209-94e1-fe80d4914377\") " pod="openstack/kube-state-metrics-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.571917 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6c69a90c-bccf-4209-94e1-fe80d4914377-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6c69a90c-bccf-4209-94e1-fe80d4914377\") " pod="openstack/kube-state-metrics-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.571980 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.571996 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.572005 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.571850 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-kube-api-access-z6bmw" (OuterVolumeSpecName: "kube-api-access-z6bmw") pod "bb1cb635-71ed-4d6f-89ba-b2dd9342a55a" (UID: "bb1cb635-71ed-4d6f-89ba-b2dd9342a55a"). InnerVolumeSpecName "kube-api-access-z6bmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.577173 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/6c69a90c-bccf-4209-94e1-fe80d4914377-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"6c69a90c-bccf-4209-94e1-fe80d4914377\") " pod="openstack/kube-state-metrics-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.581128 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c69a90c-bccf-4209-94e1-fe80d4914377-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"6c69a90c-bccf-4209-94e1-fe80d4914377\") " pod="openstack/kube-state-metrics-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.581684 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c69a90c-bccf-4209-94e1-fe80d4914377-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"6c69a90c-bccf-4209-94e1-fe80d4914377\") " pod="openstack/kube-state-metrics-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.594417 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfld7\" (UniqueName: \"kubernetes.io/projected/6c69a90c-bccf-4209-94e1-fe80d4914377-kube-api-access-pfld7\") pod \"kube-state-metrics-0\" (UID: \"6c69a90c-bccf-4209-94e1-fe80d4914377\") " pod="openstack/kube-state-metrics-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.657072 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.679877 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6bmw\" (UniqueName: \"kubernetes.io/projected/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a-kube-api-access-z6bmw\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.746246 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.781532 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74cjs\" (UniqueName: \"kubernetes.io/projected/2373439b-a619-472d-9988-e01918981b4d-kube-api-access-74cjs\") pod \"2373439b-a619-472d-9988-e01918981b4d\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.781586 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2373439b-a619-472d-9988-e01918981b4d-combined-ca-bundle\") pod \"2373439b-a619-472d-9988-e01918981b4d\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.781683 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2373439b-a619-472d-9988-e01918981b4d-config-data\") pod \"2373439b-a619-472d-9988-e01918981b4d\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.781847 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2373439b-a619-472d-9988-e01918981b4d-logs\") pod \"2373439b-a619-472d-9988-e01918981b4d\" (UID: \"2373439b-a619-472d-9988-e01918981b4d\") " Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.787136 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2373439b-a619-472d-9988-e01918981b4d-logs" (OuterVolumeSpecName: "logs") pod "2373439b-a619-472d-9988-e01918981b4d" (UID: "2373439b-a619-472d-9988-e01918981b4d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.796527 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2373439b-a619-472d-9988-e01918981b4d-kube-api-access-74cjs" (OuterVolumeSpecName: "kube-api-access-74cjs") pod "2373439b-a619-472d-9988-e01918981b4d" (UID: "2373439b-a619-472d-9988-e01918981b4d"). InnerVolumeSpecName "kube-api-access-74cjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.818963 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2373439b-a619-472d-9988-e01918981b4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2373439b-a619-472d-9988-e01918981b4d" (UID: "2373439b-a619-472d-9988-e01918981b4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.835019 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2373439b-a619-472d-9988-e01918981b4d-config-data" (OuterVolumeSpecName: "config-data") pod "2373439b-a619-472d-9988-e01918981b4d" (UID: "2373439b-a619-472d-9988-e01918981b4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.853833 4744 generic.go:334] "Generic (PLEG): container finished" podID="2373439b-a619-472d-9988-e01918981b4d" containerID="e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e" exitCode=0 Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.853867 4744 generic.go:334] "Generic (PLEG): container finished" podID="2373439b-a619-472d-9988-e01918981b4d" containerID="351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713" exitCode=143 Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.853918 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2373439b-a619-472d-9988-e01918981b4d","Type":"ContainerDied","Data":"e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e"} Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.853945 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2373439b-a619-472d-9988-e01918981b4d","Type":"ContainerDied","Data":"351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713"} Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.853955 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2373439b-a619-472d-9988-e01918981b4d","Type":"ContainerDied","Data":"fc8abf325495881da3205a82aee0ba07448743113a2290b280ac9a932fe8b102"} Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.853980 4744 scope.go:117] "RemoveContainer" containerID="e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.854117 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.875811 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qhgd6" event={"ID":"bb1cb635-71ed-4d6f-89ba-b2dd9342a55a","Type":"ContainerDied","Data":"df17f1e676ef4b7965a95b6a53a3f7781a64362021ff4bf4973205008cf20db8"} Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.875877 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df17f1e676ef4b7965a95b6a53a3f7781a64362021ff4bf4973205008cf20db8" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.876019 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qhgd6" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.894550 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2373439b-a619-472d-9988-e01918981b4d-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.894581 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74cjs\" (UniqueName: \"kubernetes.io/projected/2373439b-a619-472d-9988-e01918981b4d-kube-api-access-74cjs\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.894594 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2373439b-a619-472d-9988-e01918981b4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.894604 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2373439b-a619-472d-9988-e01918981b4d-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.904711 4744 scope.go:117] "RemoveContainer" containerID="351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.905012 4744 generic.go:334] "Generic (PLEG): container finished" podID="686a87c3-18f7-4578-a07c-ff9fb5a766cd" containerID="53872c77e10672dc21a6ea541bf521cc51f9325fd6745aca8761d4721a9c4097" exitCode=0 Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.958253 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e104e42e-cc57-4233-b7a5-2755705d18f7" path="/var/lib/kubelet/pods/e104e42e-cc57-4233-b7a5-2755705d18f7/volumes" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.960304 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.960334 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.960351 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" event={"ID":"686a87c3-18f7-4578-a07c-ff9fb5a766cd","Type":"ContainerDied","Data":"53872c77e10672dc21a6ea541bf521cc51f9325fd6745aca8761d4721a9c4097"} Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.967737 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 16:46:06 crc kubenswrapper[4744]: E1003 16:46:06.968226 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2373439b-a619-472d-9988-e01918981b4d" containerName="nova-api-log" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.968243 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2373439b-a619-472d-9988-e01918981b4d" containerName="nova-api-log" Oct 03 16:46:06 crc kubenswrapper[4744]: E1003 16:46:06.968270 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2373439b-a619-472d-9988-e01918981b4d" containerName="nova-api-api" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.968277 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2373439b-a619-472d-9988-e01918981b4d" containerName="nova-api-api" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.968445 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2373439b-a619-472d-9988-e01918981b4d" containerName="nova-api-api" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.968476 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2373439b-a619-472d-9988-e01918981b4d" containerName="nova-api-log" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.969375 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.971439 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.974574 4744 scope.go:117] "RemoveContainer" containerID="e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e" Oct 03 16:46:06 crc kubenswrapper[4744]: E1003 16:46:06.975061 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e\": container with ID starting with e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e not found: ID does not exist" containerID="e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.975099 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e"} err="failed to get container status \"e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e\": rpc error: code = NotFound desc = could not find container \"e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e\": container with ID starting with e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e not found: ID does not exist" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.975123 4744 scope.go:117] "RemoveContainer" containerID="351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713" Oct 03 16:46:06 crc kubenswrapper[4744]: E1003 16:46:06.975519 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713\": container with ID starting with 351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713 not found: ID does not exist" containerID="351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.975577 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713"} err="failed to get container status \"351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713\": rpc error: code = NotFound desc = could not find container \"351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713\": container with ID starting with 351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713 not found: ID does not exist" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.975604 4744 scope.go:117] "RemoveContainer" containerID="e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.977032 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e"} err="failed to get container status \"e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e\": rpc error: code = NotFound desc = could not find container \"e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e\": container with ID starting with e4fdc8dd3c1e8405fae1866eb66fd7a46d507a19cd969c5253a362cd77cffb8e not found: ID does not exist" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.977057 4744 scope.go:117] "RemoveContainer" containerID="351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.977115 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.978817 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.979470 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713"} err="failed to get container status \"351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713\": rpc error: code = NotFound desc = could not find container \"351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713\": container with ID starting with 351fae2d7926ea3964471c70319296b10b6152dc62faf08856448799130ff713 not found: ID does not exist" Oct 03 16:46:06 crc kubenswrapper[4744]: I1003 16:46:06.984622 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.001446 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.019486 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.042849 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.097764 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-ovsdbserver-nb\") pod \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.097899 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-ovsdbserver-sb\") pod \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.098001 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5qbd\" (UniqueName: \"kubernetes.io/projected/686a87c3-18f7-4578-a07c-ff9fb5a766cd-kube-api-access-g5qbd\") pod \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.098049 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-dns-svc\") pod \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.098159 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-config\") pod \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.098260 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-dns-swift-storage-0\") pod \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\" (UID: \"686a87c3-18f7-4578-a07c-ff9fb5a766cd\") " Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.098560 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06c6c196-5109-4edb-a034-bd5d45cb5873-config-data\") pod \"nova-api-0\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " pod="openstack/nova-api-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.098602 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qg2q\" (UniqueName: \"kubernetes.io/projected/06c6c196-5109-4edb-a034-bd5d45cb5873-kube-api-access-5qg2q\") pod \"nova-api-0\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " pod="openstack/nova-api-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.098633 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a40f5d-b401-437e-9e4c-38e1d3459e40-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"74a40f5d-b401-437e-9e4c-38e1d3459e40\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.098779 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06c6c196-5109-4edb-a034-bd5d45cb5873-logs\") pod \"nova-api-0\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " pod="openstack/nova-api-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.098834 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a40f5d-b401-437e-9e4c-38e1d3459e40-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"74a40f5d-b401-437e-9e4c-38e1d3459e40\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.098867 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw2w8\" (UniqueName: \"kubernetes.io/projected/74a40f5d-b401-437e-9e4c-38e1d3459e40-kube-api-access-dw2w8\") pod \"nova-cell1-conductor-0\" (UID: \"74a40f5d-b401-437e-9e4c-38e1d3459e40\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.098990 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c6c196-5109-4edb-a034-bd5d45cb5873-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " pod="openstack/nova-api-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.108730 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/686a87c3-18f7-4578-a07c-ff9fb5a766cd-kube-api-access-g5qbd" (OuterVolumeSpecName: "kube-api-access-g5qbd") pod "686a87c3-18f7-4578-a07c-ff9fb5a766cd" (UID: "686a87c3-18f7-4578-a07c-ff9fb5a766cd"). InnerVolumeSpecName "kube-api-access-g5qbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.156034 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "686a87c3-18f7-4578-a07c-ff9fb5a766cd" (UID: "686a87c3-18f7-4578-a07c-ff9fb5a766cd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.169170 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "686a87c3-18f7-4578-a07c-ff9fb5a766cd" (UID: "686a87c3-18f7-4578-a07c-ff9fb5a766cd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.176997 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-config" (OuterVolumeSpecName: "config") pod "686a87c3-18f7-4578-a07c-ff9fb5a766cd" (UID: "686a87c3-18f7-4578-a07c-ff9fb5a766cd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.187641 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "686a87c3-18f7-4578-a07c-ff9fb5a766cd" (UID: "686a87c3-18f7-4578-a07c-ff9fb5a766cd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.198733 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "686a87c3-18f7-4578-a07c-ff9fb5a766cd" (UID: "686a87c3-18f7-4578-a07c-ff9fb5a766cd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.200833 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a40f5d-b401-437e-9e4c-38e1d3459e40-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"74a40f5d-b401-437e-9e4c-38e1d3459e40\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.200989 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06c6c196-5109-4edb-a034-bd5d45cb5873-logs\") pod \"nova-api-0\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " pod="openstack/nova-api-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.201028 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a40f5d-b401-437e-9e4c-38e1d3459e40-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"74a40f5d-b401-437e-9e4c-38e1d3459e40\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.201052 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw2w8\" (UniqueName: \"kubernetes.io/projected/74a40f5d-b401-437e-9e4c-38e1d3459e40-kube-api-access-dw2w8\") pod \"nova-cell1-conductor-0\" (UID: \"74a40f5d-b401-437e-9e4c-38e1d3459e40\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.201091 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c6c196-5109-4edb-a034-bd5d45cb5873-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " pod="openstack/nova-api-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.201126 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06c6c196-5109-4edb-a034-bd5d45cb5873-config-data\") pod \"nova-api-0\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " pod="openstack/nova-api-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.201147 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qg2q\" (UniqueName: \"kubernetes.io/projected/06c6c196-5109-4edb-a034-bd5d45cb5873-kube-api-access-5qg2q\") pod \"nova-api-0\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " pod="openstack/nova-api-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.201209 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.201225 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.201236 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.201245 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.201254 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5qbd\" (UniqueName: \"kubernetes.io/projected/686a87c3-18f7-4578-a07c-ff9fb5a766cd-kube-api-access-g5qbd\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.201263 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/686a87c3-18f7-4578-a07c-ff9fb5a766cd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.201393 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06c6c196-5109-4edb-a034-bd5d45cb5873-logs\") pod \"nova-api-0\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " pod="openstack/nova-api-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.216183 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c6c196-5109-4edb-a034-bd5d45cb5873-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " pod="openstack/nova-api-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.216269 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a40f5d-b401-437e-9e4c-38e1d3459e40-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"74a40f5d-b401-437e-9e4c-38e1d3459e40\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.218560 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a40f5d-b401-437e-9e4c-38e1d3459e40-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"74a40f5d-b401-437e-9e4c-38e1d3459e40\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.222783 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06c6c196-5109-4edb-a034-bd5d45cb5873-config-data\") pod \"nova-api-0\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " pod="openstack/nova-api-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.223209 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw2w8\" (UniqueName: \"kubernetes.io/projected/74a40f5d-b401-437e-9e4c-38e1d3459e40-kube-api-access-dw2w8\") pod \"nova-cell1-conductor-0\" (UID: \"74a40f5d-b401-437e-9e4c-38e1d3459e40\") " pod="openstack/nova-cell1-conductor-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.223239 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qg2q\" (UniqueName: \"kubernetes.io/projected/06c6c196-5109-4edb-a034-bd5d45cb5873-kube-api-access-5qg2q\") pod \"nova-api-0\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " pod="openstack/nova-api-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.286812 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 16:46:07 crc kubenswrapper[4744]: W1003 16:46:07.314359 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c69a90c_bccf_4209_94e1_fe80d4914377.slice/crio-2ab5acec645be8b9082dc2a637845974a5e5469da639dc3b8ee0e2108458190a WatchSource:0}: Error finding container 2ab5acec645be8b9082dc2a637845974a5e5469da639dc3b8ee0e2108458190a: Status 404 returned error can't find the container with id 2ab5acec645be8b9082dc2a637845974a5e5469da639dc3b8ee0e2108458190a Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.318773 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.318990 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.422834 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.423269 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="ceilometer-central-agent" containerID="cri-o://7e1b1054724271bd934c3839f03d00046bc3cdc126777b683d26b07640b0c338" gracePeriod=30 Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.423814 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="proxy-httpd" containerID="cri-o://34c8ada08e7d738040836f8abe9db2483c2abe4990e9c67d0804fb55756080ef" gracePeriod=30 Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.426151 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="ceilometer-notification-agent" containerID="cri-o://503df61f5a4417b67b6bec262434d92ed9650d3845139b4ebbe0d4dcd3706c01" gracePeriod=30 Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.426187 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="sg-core" containerID="cri-o://58a8d3a5b2ee319eae5b88614209bd8af5fcdef4f06231b1770f1a2be95c566f" gracePeriod=30 Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.746503 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 16:46:07 crc kubenswrapper[4744]: W1003 16:46:07.778785 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74a40f5d_b401_437e_9e4c_38e1d3459e40.slice/crio-8791721d46a1c910d31af60ee0b6ae61ecdd118558bb43620e38647b89cb195c WatchSource:0}: Error finding container 8791721d46a1c910d31af60ee0b6ae61ecdd118558bb43620e38647b89cb195c: Status 404 returned error can't find the container with id 8791721d46a1c910d31af60ee0b6ae61ecdd118558bb43620e38647b89cb195c Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.867623 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:07 crc kubenswrapper[4744]: W1003 16:46:07.881454 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06c6c196_5109_4edb_a034_bd5d45cb5873.slice/crio-95c35046d05dff5b9b6440c91e4012edf4cfd208c48b46ce2d3d52a950be4c5b WatchSource:0}: Error finding container 95c35046d05dff5b9b6440c91e4012edf4cfd208c48b46ce2d3d52a950be4c5b: Status 404 returned error can't find the container with id 95c35046d05dff5b9b6440c91e4012edf4cfd208c48b46ce2d3d52a950be4c5b Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.934792 4744 generic.go:334] "Generic (PLEG): container finished" podID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerID="34c8ada08e7d738040836f8abe9db2483c2abe4990e9c67d0804fb55756080ef" exitCode=0 Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.935160 4744 generic.go:334] "Generic (PLEG): container finished" podID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerID="58a8d3a5b2ee319eae5b88614209bd8af5fcdef4f06231b1770f1a2be95c566f" exitCode=2 Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.935173 4744 generic.go:334] "Generic (PLEG): container finished" podID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerID="7e1b1054724271bd934c3839f03d00046bc3cdc126777b683d26b07640b0c338" exitCode=0 Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.934873 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d74ef7e-3629-4819-a3d8-a8f7fd49b284","Type":"ContainerDied","Data":"34c8ada08e7d738040836f8abe9db2483c2abe4990e9c67d0804fb55756080ef"} Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.935261 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d74ef7e-3629-4819-a3d8-a8f7fd49b284","Type":"ContainerDied","Data":"58a8d3a5b2ee319eae5b88614209bd8af5fcdef4f06231b1770f1a2be95c566f"} Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.935274 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d74ef7e-3629-4819-a3d8-a8f7fd49b284","Type":"ContainerDied","Data":"7e1b1054724271bd934c3839f03d00046bc3cdc126777b683d26b07640b0c338"} Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.938273 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"74a40f5d-b401-437e-9e4c-38e1d3459e40","Type":"ContainerStarted","Data":"8791721d46a1c910d31af60ee0b6ae61ecdd118558bb43620e38647b89cb195c"} Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.943517 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.943537 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5865f9d689-cwtl8" event={"ID":"686a87c3-18f7-4578-a07c-ff9fb5a766cd","Type":"ContainerDied","Data":"2293af597a99ea310d60af102b24a55a0d7fe2e8d4568c1d695276438bfdab28"} Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.943596 4744 scope.go:117] "RemoveContainer" containerID="53872c77e10672dc21a6ea541bf521cc51f9325fd6745aca8761d4721a9c4097" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.952560 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6c69a90c-bccf-4209-94e1-fe80d4914377","Type":"ContainerStarted","Data":"2ab5acec645be8b9082dc2a637845974a5e5469da639dc3b8ee0e2108458190a"} Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.954590 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"06c6c196-5109-4edb-a034-bd5d45cb5873","Type":"ContainerStarted","Data":"95c35046d05dff5b9b6440c91e4012edf4cfd208c48b46ce2d3d52a950be4c5b"} Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.958393 4744 generic.go:334] "Generic (PLEG): container finished" podID="133d6831-f643-4e6f-b245-ce4cca4a1b8c" containerID="922b19d530b25702f80d628149b1977891e6a4399d41d708e3cfd9744f7dfa22" exitCode=0 Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.958431 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"133d6831-f643-4e6f-b245-ce4cca4a1b8c","Type":"ContainerDied","Data":"922b19d530b25702f80d628149b1977891e6a4399d41d708e3cfd9744f7dfa22"} Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.977166 4744 scope.go:117] "RemoveContainer" containerID="f280fc967e97cb2fff2a4a3b560febe1c9455a1f3ff75052ded866b0a151aee6" Oct 03 16:46:07 crc kubenswrapper[4744]: I1003 16:46:07.996244 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5865f9d689-cwtl8"] Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.014269 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5865f9d689-cwtl8"] Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.126546 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.220685 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133d6831-f643-4e6f-b245-ce4cca4a1b8c-config-data\") pod \"133d6831-f643-4e6f-b245-ce4cca4a1b8c\" (UID: \"133d6831-f643-4e6f-b245-ce4cca4a1b8c\") " Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.220806 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133d6831-f643-4e6f-b245-ce4cca4a1b8c-combined-ca-bundle\") pod \"133d6831-f643-4e6f-b245-ce4cca4a1b8c\" (UID: \"133d6831-f643-4e6f-b245-ce4cca4a1b8c\") " Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.220829 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgvc8\" (UniqueName: \"kubernetes.io/projected/133d6831-f643-4e6f-b245-ce4cca4a1b8c-kube-api-access-tgvc8\") pod \"133d6831-f643-4e6f-b245-ce4cca4a1b8c\" (UID: \"133d6831-f643-4e6f-b245-ce4cca4a1b8c\") " Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.226064 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/133d6831-f643-4e6f-b245-ce4cca4a1b8c-kube-api-access-tgvc8" (OuterVolumeSpecName: "kube-api-access-tgvc8") pod "133d6831-f643-4e6f-b245-ce4cca4a1b8c" (UID: "133d6831-f643-4e6f-b245-ce4cca4a1b8c"). InnerVolumeSpecName "kube-api-access-tgvc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.251394 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/133d6831-f643-4e6f-b245-ce4cca4a1b8c-config-data" (OuterVolumeSpecName: "config-data") pod "133d6831-f643-4e6f-b245-ce4cca4a1b8c" (UID: "133d6831-f643-4e6f-b245-ce4cca4a1b8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.268858 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/133d6831-f643-4e6f-b245-ce4cca4a1b8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "133d6831-f643-4e6f-b245-ce4cca4a1b8c" (UID: "133d6831-f643-4e6f-b245-ce4cca4a1b8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.323417 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/133d6831-f643-4e6f-b245-ce4cca4a1b8c-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.323449 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/133d6831-f643-4e6f-b245-ce4cca4a1b8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.323464 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgvc8\" (UniqueName: \"kubernetes.io/projected/133d6831-f643-4e6f-b245-ce4cca4a1b8c-kube-api-access-tgvc8\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.907014 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2373439b-a619-472d-9988-e01918981b4d" path="/var/lib/kubelet/pods/2373439b-a619-472d-9988-e01918981b4d/volumes" Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.907889 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="686a87c3-18f7-4578-a07c-ff9fb5a766cd" path="/var/lib/kubelet/pods/686a87c3-18f7-4578-a07c-ff9fb5a766cd/volumes" Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.970832 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6c69a90c-bccf-4209-94e1-fe80d4914377","Type":"ContainerStarted","Data":"d6eb4aadea61e8c09f194a83f1f623a5032aecdc2f802b6a52b4b0446f237f2d"} Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.972282 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.975341 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"06c6c196-5109-4edb-a034-bd5d45cb5873","Type":"ContainerStarted","Data":"5f8b355b7009ffa37cb32fe8190cf0452a22787db1bc87dec02ef3d70ca2f80f"} Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.975377 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"06c6c196-5109-4edb-a034-bd5d45cb5873","Type":"ContainerStarted","Data":"597bdf8f502be320b1f3325fad9cc3ddffe4bfc96b395d70daec911a1c87e9cb"} Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.978229 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"133d6831-f643-4e6f-b245-ce4cca4a1b8c","Type":"ContainerDied","Data":"3dc654619212b2f17f51aab006ee84edc790af1d32f5406cd4de5d5594355f79"} Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.978267 4744 scope.go:117] "RemoveContainer" containerID="922b19d530b25702f80d628149b1977891e6a4399d41d708e3cfd9744f7dfa22" Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.978379 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.985285 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"74a40f5d-b401-437e-9e4c-38e1d3459e40","Type":"ContainerStarted","Data":"01cdba7a49caa0304567fa96eaed9ef365fb22d7c838ca1c5ecd30025053da83"} Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.985592 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 03 16:46:08 crc kubenswrapper[4744]: I1003 16:46:08.994868 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.60744735 podStartE2EDuration="2.994846988s" podCreationTimestamp="2025-10-03 16:46:06 +0000 UTC" firstStartedPulling="2025-10-03 16:46:07.317832074 +0000 UTC m=+1293.597707970" lastFinishedPulling="2025-10-03 16:46:07.705231712 +0000 UTC m=+1293.985107608" observedRunningTime="2025-10-03 16:46:08.989682466 +0000 UTC m=+1295.269558382" watchObservedRunningTime="2025-10-03 16:46:08.994846988 +0000 UTC m=+1295.274722884" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.012643 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.01262482 podStartE2EDuration="3.01262482s" podCreationTimestamp="2025-10-03 16:46:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:46:09.008283129 +0000 UTC m=+1295.288159045" watchObservedRunningTime="2025-10-03 16:46:09.01262482 +0000 UTC m=+1295.292500716" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.034260 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.034235399 podStartE2EDuration="3.034235399s" podCreationTimestamp="2025-10-03 16:46:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:46:09.026081262 +0000 UTC m=+1295.305957188" watchObservedRunningTime="2025-10-03 16:46:09.034235399 +0000 UTC m=+1295.314111315" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.057346 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.063147 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.071196 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:46:09 crc kubenswrapper[4744]: E1003 16:46:09.071776 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="686a87c3-18f7-4578-a07c-ff9fb5a766cd" containerName="dnsmasq-dns" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.071796 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="686a87c3-18f7-4578-a07c-ff9fb5a766cd" containerName="dnsmasq-dns" Oct 03 16:46:09 crc kubenswrapper[4744]: E1003 16:46:09.071818 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="133d6831-f643-4e6f-b245-ce4cca4a1b8c" containerName="nova-scheduler-scheduler" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.071826 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="133d6831-f643-4e6f-b245-ce4cca4a1b8c" containerName="nova-scheduler-scheduler" Oct 03 16:46:09 crc kubenswrapper[4744]: E1003 16:46:09.071867 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="686a87c3-18f7-4578-a07c-ff9fb5a766cd" containerName="init" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.071874 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="686a87c3-18f7-4578-a07c-ff9fb5a766cd" containerName="init" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.072117 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="686a87c3-18f7-4578-a07c-ff9fb5a766cd" containerName="dnsmasq-dns" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.072141 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="133d6831-f643-4e6f-b245-ce4cca4a1b8c" containerName="nova-scheduler-scheduler" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.072887 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.075599 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.083600 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.139764 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3705e0e-22b4-4691-b190-4173ff857f90-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d3705e0e-22b4-4691-b190-4173ff857f90\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.139817 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ctdr\" (UniqueName: \"kubernetes.io/projected/d3705e0e-22b4-4691-b190-4173ff857f90-kube-api-access-9ctdr\") pod \"nova-scheduler-0\" (UID: \"d3705e0e-22b4-4691-b190-4173ff857f90\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.139860 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3705e0e-22b4-4691-b190-4173ff857f90-config-data\") pod \"nova-scheduler-0\" (UID: \"d3705e0e-22b4-4691-b190-4173ff857f90\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.241256 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3705e0e-22b4-4691-b190-4173ff857f90-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d3705e0e-22b4-4691-b190-4173ff857f90\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.241303 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ctdr\" (UniqueName: \"kubernetes.io/projected/d3705e0e-22b4-4691-b190-4173ff857f90-kube-api-access-9ctdr\") pod \"nova-scheduler-0\" (UID: \"d3705e0e-22b4-4691-b190-4173ff857f90\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.241338 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3705e0e-22b4-4691-b190-4173ff857f90-config-data\") pod \"nova-scheduler-0\" (UID: \"d3705e0e-22b4-4691-b190-4173ff857f90\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.253321 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3705e0e-22b4-4691-b190-4173ff857f90-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d3705e0e-22b4-4691-b190-4173ff857f90\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.262125 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3705e0e-22b4-4691-b190-4173ff857f90-config-data\") pod \"nova-scheduler-0\" (UID: \"d3705e0e-22b4-4691-b190-4173ff857f90\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.264602 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ctdr\" (UniqueName: \"kubernetes.io/projected/d3705e0e-22b4-4691-b190-4173ff857f90-kube-api-access-9ctdr\") pod \"nova-scheduler-0\" (UID: \"d3705e0e-22b4-4691-b190-4173ff857f90\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.394203 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.932664 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:46:09 crc kubenswrapper[4744]: I1003 16:46:09.997233 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d3705e0e-22b4-4691-b190-4173ff857f90","Type":"ContainerStarted","Data":"3b7aa633beb22d32f429b98381c9b4d4f1834573a4bcab3c9ace4a65ddde055d"} Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.001547 4744 generic.go:334] "Generic (PLEG): container finished" podID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerID="503df61f5a4417b67b6bec262434d92ed9650d3845139b4ebbe0d4dcd3706c01" exitCode=0 Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.001662 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d74ef7e-3629-4819-a3d8-a8f7fd49b284","Type":"ContainerDied","Data":"503df61f5a4417b67b6bec262434d92ed9650d3845139b4ebbe0d4dcd3706c01"} Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.020751 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.059198 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-combined-ca-bundle\") pod \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.059256 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-log-httpd\") pod \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.059325 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-sg-core-conf-yaml\") pod \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.059348 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-config-data\") pod \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.059373 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq9gk\" (UniqueName: \"kubernetes.io/projected/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-kube-api-access-lq9gk\") pod \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.059437 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-scripts\") pod \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.059470 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-run-httpd\") pod \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\" (UID: \"5d74ef7e-3629-4819-a3d8-a8f7fd49b284\") " Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.060258 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5d74ef7e-3629-4819-a3d8-a8f7fd49b284" (UID: "5d74ef7e-3629-4819-a3d8-a8f7fd49b284"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.064072 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5d74ef7e-3629-4819-a3d8-a8f7fd49b284" (UID: "5d74ef7e-3629-4819-a3d8-a8f7fd49b284"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.067108 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-kube-api-access-lq9gk" (OuterVolumeSpecName: "kube-api-access-lq9gk") pod "5d74ef7e-3629-4819-a3d8-a8f7fd49b284" (UID: "5d74ef7e-3629-4819-a3d8-a8f7fd49b284"). InnerVolumeSpecName "kube-api-access-lq9gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.076993 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-scripts" (OuterVolumeSpecName: "scripts") pod "5d74ef7e-3629-4819-a3d8-a8f7fd49b284" (UID: "5d74ef7e-3629-4819-a3d8-a8f7fd49b284"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.109739 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5d74ef7e-3629-4819-a3d8-a8f7fd49b284" (UID: "5d74ef7e-3629-4819-a3d8-a8f7fd49b284"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.161642 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.161675 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.161687 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.161698 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.161709 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq9gk\" (UniqueName: \"kubernetes.io/projected/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-kube-api-access-lq9gk\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.176550 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d74ef7e-3629-4819-a3d8-a8f7fd49b284" (UID: "5d74ef7e-3629-4819-a3d8-a8f7fd49b284"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.201957 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-config-data" (OuterVolumeSpecName: "config-data") pod "5d74ef7e-3629-4819-a3d8-a8f7fd49b284" (UID: "5d74ef7e-3629-4819-a3d8-a8f7fd49b284"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.263919 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.263962 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d74ef7e-3629-4819-a3d8-a8f7fd49b284-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:10 crc kubenswrapper[4744]: I1003 16:46:10.903468 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="133d6831-f643-4e6f-b245-ce4cca4a1b8c" path="/var/lib/kubelet/pods/133d6831-f643-4e6f-b245-ce4cca4a1b8c/volumes" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.011327 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d3705e0e-22b4-4691-b190-4173ff857f90","Type":"ContainerStarted","Data":"0962d98fee0e2b6d711d5b4cf85c5ab78c7ef9fa77c484483c28a583ac101230"} Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.015564 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.016129 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5d74ef7e-3629-4819-a3d8-a8f7fd49b284","Type":"ContainerDied","Data":"ec4ec5c6ad26b9832aaec6cab83972d68b3e643f27fd30ee94cf15644d4d0d0a"} Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.016165 4744 scope.go:117] "RemoveContainer" containerID="34c8ada08e7d738040836f8abe9db2483c2abe4990e9c67d0804fb55756080ef" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.039696 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.039676703 podStartE2EDuration="2.039676703s" podCreationTimestamp="2025-10-03 16:46:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:46:11.036611685 +0000 UTC m=+1297.316487581" watchObservedRunningTime="2025-10-03 16:46:11.039676703 +0000 UTC m=+1297.319552599" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.040911 4744 scope.go:117] "RemoveContainer" containerID="58a8d3a5b2ee319eae5b88614209bd8af5fcdef4f06231b1770f1a2be95c566f" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.063414 4744 scope.go:117] "RemoveContainer" containerID="503df61f5a4417b67b6bec262434d92ed9650d3845139b4ebbe0d4dcd3706c01" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.063587 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.101459 4744 scope.go:117] "RemoveContainer" containerID="7e1b1054724271bd934c3839f03d00046bc3cdc126777b683d26b07640b0c338" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.104615 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.126667 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:46:11 crc kubenswrapper[4744]: E1003 16:46:11.127112 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="ceilometer-central-agent" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.127130 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="ceilometer-central-agent" Oct 03 16:46:11 crc kubenswrapper[4744]: E1003 16:46:11.127147 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="proxy-httpd" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.127154 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="proxy-httpd" Oct 03 16:46:11 crc kubenswrapper[4744]: E1003 16:46:11.127173 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="ceilometer-notification-agent" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.127179 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="ceilometer-notification-agent" Oct 03 16:46:11 crc kubenswrapper[4744]: E1003 16:46:11.127201 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="sg-core" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.127207 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="sg-core" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.127376 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="proxy-httpd" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.127386 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="sg-core" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.127405 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="ceilometer-notification-agent" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.127419 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" containerName="ceilometer-central-agent" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.129452 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.131447 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.131642 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.136418 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.140282 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.282093 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xq4d\" (UniqueName: \"kubernetes.io/projected/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-kube-api-access-6xq4d\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.282147 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.282186 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-config-data\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.282257 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.282290 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-log-httpd\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.282312 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-run-httpd\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.282343 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.282359 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-scripts\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.383895 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xq4d\" (UniqueName: \"kubernetes.io/projected/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-kube-api-access-6xq4d\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.383943 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.383982 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-config-data\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.384031 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.384061 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-log-httpd\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.384082 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-run-httpd\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.384114 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.384130 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-scripts\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.385013 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-run-httpd\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.385233 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-log-httpd\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.389925 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.389981 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.390020 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-scripts\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.390670 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-config-data\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.391035 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.414992 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xq4d\" (UniqueName: \"kubernetes.io/projected/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-kube-api-access-6xq4d\") pod \"ceilometer-0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.453223 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:46:11 crc kubenswrapper[4744]: I1003 16:46:11.906658 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:46:11 crc kubenswrapper[4744]: W1003 16:46:11.918110 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d39052e_6ed5_47a7_9e91_2b47f55d94b0.slice/crio-a4ae2fc86dd093fea2daf39eb908eaecf4a3b15710ec846627d68cefa5df47d6 WatchSource:0}: Error finding container a4ae2fc86dd093fea2daf39eb908eaecf4a3b15710ec846627d68cefa5df47d6: Status 404 returned error can't find the container with id a4ae2fc86dd093fea2daf39eb908eaecf4a3b15710ec846627d68cefa5df47d6 Oct 03 16:46:12 crc kubenswrapper[4744]: I1003 16:46:12.033266 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d39052e-6ed5-47a7-9e91-2b47f55d94b0","Type":"ContainerStarted","Data":"a4ae2fc86dd093fea2daf39eb908eaecf4a3b15710ec846627d68cefa5df47d6"} Oct 03 16:46:12 crc kubenswrapper[4744]: I1003 16:46:12.905682 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d74ef7e-3629-4819-a3d8-a8f7fd49b284" path="/var/lib/kubelet/pods/5d74ef7e-3629-4819-a3d8-a8f7fd49b284/volumes" Oct 03 16:46:13 crc kubenswrapper[4744]: I1003 16:46:13.050133 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d39052e-6ed5-47a7-9e91-2b47f55d94b0","Type":"ContainerStarted","Data":"405b4de4b9ea0cdd5e0dc6d254303bc0cac67c370ae38300acbf1e24cbd4b002"} Oct 03 16:46:14 crc kubenswrapper[4744]: I1003 16:46:14.060754 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d39052e-6ed5-47a7-9e91-2b47f55d94b0","Type":"ContainerStarted","Data":"6773f466de0370c325a944eb30b9da846e0dbbdd3cd8264d7000bda0b6b1680a"} Oct 03 16:46:14 crc kubenswrapper[4744]: I1003 16:46:14.395352 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 03 16:46:15 crc kubenswrapper[4744]: I1003 16:46:15.073890 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d39052e-6ed5-47a7-9e91-2b47f55d94b0","Type":"ContainerStarted","Data":"9cd2c08fb236128fe62c1bd88e4109a7f4ac2235f7e5995c66fe8d10d7b7a10e"} Oct 03 16:46:16 crc kubenswrapper[4744]: I1003 16:46:16.087283 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d39052e-6ed5-47a7-9e91-2b47f55d94b0","Type":"ContainerStarted","Data":"22c7ed5bc212ba6d0f0fd30d281e9f733ddda1db86129a7f3690833e47169b41"} Oct 03 16:46:16 crc kubenswrapper[4744]: I1003 16:46:16.087413 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 16:46:16 crc kubenswrapper[4744]: I1003 16:46:16.124408 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.274366831 podStartE2EDuration="5.124381839s" podCreationTimestamp="2025-10-03 16:46:11 +0000 UTC" firstStartedPulling="2025-10-03 16:46:11.922712831 +0000 UTC m=+1298.202588737" lastFinishedPulling="2025-10-03 16:46:15.772727849 +0000 UTC m=+1302.052603745" observedRunningTime="2025-10-03 16:46:16.116890188 +0000 UTC m=+1302.396766084" watchObservedRunningTime="2025-10-03 16:46:16.124381839 +0000 UTC m=+1302.404257745" Oct 03 16:46:16 crc kubenswrapper[4744]: I1003 16:46:16.753131 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 03 16:46:17 crc kubenswrapper[4744]: I1003 16:46:17.315885 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 03 16:46:17 crc kubenswrapper[4744]: I1003 16:46:17.319474 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 16:46:17 crc kubenswrapper[4744]: I1003 16:46:17.319605 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 16:46:18 crc kubenswrapper[4744]: I1003 16:46:18.403658 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="06c6c196-5109-4edb-a034-bd5d45cb5873" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.207:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 16:46:18 crc kubenswrapper[4744]: I1003 16:46:18.403658 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="06c6c196-5109-4edb-a034-bd5d45cb5873" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.207:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 16:46:19 crc kubenswrapper[4744]: I1003 16:46:19.395422 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 03 16:46:19 crc kubenswrapper[4744]: I1003 16:46:19.437022 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 03 16:46:20 crc kubenswrapper[4744]: I1003 16:46:20.148817 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 03 16:46:27 crc kubenswrapper[4744]: I1003 16:46:27.324637 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 16:46:27 crc kubenswrapper[4744]: I1003 16:46:27.326026 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 16:46:27 crc kubenswrapper[4744]: I1003 16:46:27.326804 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 16:46:27 crc kubenswrapper[4744]: I1003 16:46:27.328775 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.224643 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.228301 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.411073 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6559f4fbd7-pfrnh"] Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.413172 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.435370 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6559f4fbd7-pfrnh"] Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.465648 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j44tw\" (UniqueName: \"kubernetes.io/projected/c9c81893-45c2-402f-8897-ed9d019f78b1-kube-api-access-j44tw\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.465715 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-ovsdbserver-nb\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.465765 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-dns-swift-storage-0\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.465815 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-config\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.465852 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-dns-svc\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.465869 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-ovsdbserver-sb\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.567850 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-ovsdbserver-nb\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.567956 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-dns-swift-storage-0\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.568023 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-config\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.568074 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-dns-svc\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.568099 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-ovsdbserver-sb\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.568187 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j44tw\" (UniqueName: \"kubernetes.io/projected/c9c81893-45c2-402f-8897-ed9d019f78b1-kube-api-access-j44tw\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.568836 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-ovsdbserver-nb\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.568924 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-config\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.569061 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-ovsdbserver-sb\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.569378 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-dns-swift-storage-0\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.569636 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-dns-svc\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.586691 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j44tw\" (UniqueName: \"kubernetes.io/projected/c9c81893-45c2-402f-8897-ed9d019f78b1-kube-api-access-j44tw\") pod \"dnsmasq-dns-6559f4fbd7-pfrnh\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:28 crc kubenswrapper[4744]: I1003 16:46:28.744201 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:29 crc kubenswrapper[4744]: I1003 16:46:29.206214 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6559f4fbd7-pfrnh"] Oct 03 16:46:29 crc kubenswrapper[4744]: W1003 16:46:29.211593 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9c81893_45c2_402f_8897_ed9d019f78b1.slice/crio-eb3636df3315e5e04417bfdc331d0837c8b44d024c9e315916ebde71e1a65eec WatchSource:0}: Error finding container eb3636df3315e5e04417bfdc331d0837c8b44d024c9e315916ebde71e1a65eec: Status 404 returned error can't find the container with id eb3636df3315e5e04417bfdc331d0837c8b44d024c9e315916ebde71e1a65eec Oct 03 16:46:29 crc kubenswrapper[4744]: I1003 16:46:29.235363 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" event={"ID":"c9c81893-45c2-402f-8897-ed9d019f78b1","Type":"ContainerStarted","Data":"eb3636df3315e5e04417bfdc331d0837c8b44d024c9e315916ebde71e1a65eec"} Oct 03 16:46:30 crc kubenswrapper[4744]: I1003 16:46:30.251653 4744 generic.go:334] "Generic (PLEG): container finished" podID="c9c81893-45c2-402f-8897-ed9d019f78b1" containerID="0f7ce8b158c59555f61798f11003a41c5d818ee2610f7c2906773f003db2bae5" exitCode=0 Oct 03 16:46:30 crc kubenswrapper[4744]: I1003 16:46:30.251759 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" event={"ID":"c9c81893-45c2-402f-8897-ed9d019f78b1","Type":"ContainerDied","Data":"0f7ce8b158c59555f61798f11003a41c5d818ee2610f7c2906773f003db2bae5"} Oct 03 16:46:30 crc kubenswrapper[4744]: I1003 16:46:30.356970 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:46:30 crc kubenswrapper[4744]: I1003 16:46:30.357285 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="ceilometer-central-agent" containerID="cri-o://405b4de4b9ea0cdd5e0dc6d254303bc0cac67c370ae38300acbf1e24cbd4b002" gracePeriod=30 Oct 03 16:46:30 crc kubenswrapper[4744]: I1003 16:46:30.357700 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="proxy-httpd" containerID="cri-o://22c7ed5bc212ba6d0f0fd30d281e9f733ddda1db86129a7f3690833e47169b41" gracePeriod=30 Oct 03 16:46:30 crc kubenswrapper[4744]: I1003 16:46:30.357882 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="ceilometer-notification-agent" containerID="cri-o://6773f466de0370c325a944eb30b9da846e0dbbdd3cd8264d7000bda0b6b1680a" gracePeriod=30 Oct 03 16:46:30 crc kubenswrapper[4744]: I1003 16:46:30.358211 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="sg-core" containerID="cri-o://9cd2c08fb236128fe62c1bd88e4109a7f4ac2235f7e5995c66fe8d10d7b7a10e" gracePeriod=30 Oct 03 16:46:30 crc kubenswrapper[4744]: I1003 16:46:30.472373 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.209:3000/\": read tcp 10.217.0.2:60802->10.217.0.209:3000: read: connection reset by peer" Oct 03 16:46:30 crc kubenswrapper[4744]: I1003 16:46:30.978996 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.261530 4744 generic.go:334] "Generic (PLEG): container finished" podID="2528f1c7-4a26-472c-84a5-c5d0a90c5ece" containerID="d6ac5af47e78703484a6564224485d108b26016a512fe43491200033e1079231" exitCode=137 Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.261603 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2528f1c7-4a26-472c-84a5-c5d0a90c5ece","Type":"ContainerDied","Data":"d6ac5af47e78703484a6564224485d108b26016a512fe43491200033e1079231"} Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.261941 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2528f1c7-4a26-472c-84a5-c5d0a90c5ece","Type":"ContainerDied","Data":"040941ba0767c5e086195037328eb12e86a407d56256cef58adeb58326778b27"} Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.261972 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="040941ba0767c5e086195037328eb12e86a407d56256cef58adeb58326778b27" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.264882 4744 generic.go:334] "Generic (PLEG): container finished" podID="ea42249b-a5a1-4728-8be6-8babf3c51aa2" containerID="2d3ec7a1b23df0aeefde51845f97da1a41107b85563c69c240d72fdd219d0625" exitCode=137 Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.264946 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ea42249b-a5a1-4728-8be6-8babf3c51aa2","Type":"ContainerDied","Data":"2d3ec7a1b23df0aeefde51845f97da1a41107b85563c69c240d72fdd219d0625"} Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.264967 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ea42249b-a5a1-4728-8be6-8babf3c51aa2","Type":"ContainerDied","Data":"e0f923a0cc6165a2f33f087670219971545518476e8d2bc29526480aec44f3d2"} Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.264981 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0f923a0cc6165a2f33f087670219971545518476e8d2bc29526480aec44f3d2" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.265011 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.266690 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" event={"ID":"c9c81893-45c2-402f-8897-ed9d019f78b1","Type":"ContainerStarted","Data":"8c1cb84ff35f8a783b54f45d869f3ba51577f5ce4d61af74f55883c8da7643ad"} Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.267687 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.270233 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerID="22c7ed5bc212ba6d0f0fd30d281e9f733ddda1db86129a7f3690833e47169b41" exitCode=0 Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.270259 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerID="9cd2c08fb236128fe62c1bd88e4109a7f4ac2235f7e5995c66fe8d10d7b7a10e" exitCode=2 Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.270270 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerID="405b4de4b9ea0cdd5e0dc6d254303bc0cac67c370ae38300acbf1e24cbd4b002" exitCode=0 Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.270299 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d39052e-6ed5-47a7-9e91-2b47f55d94b0","Type":"ContainerDied","Data":"22c7ed5bc212ba6d0f0fd30d281e9f733ddda1db86129a7f3690833e47169b41"} Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.270331 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d39052e-6ed5-47a7-9e91-2b47f55d94b0","Type":"ContainerDied","Data":"9cd2c08fb236128fe62c1bd88e4109a7f4ac2235f7e5995c66fe8d10d7b7a10e"} Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.270342 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d39052e-6ed5-47a7-9e91-2b47f55d94b0","Type":"ContainerDied","Data":"405b4de4b9ea0cdd5e0dc6d254303bc0cac67c370ae38300acbf1e24cbd4b002"} Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.270620 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="06c6c196-5109-4edb-a034-bd5d45cb5873" containerName="nova-api-api" containerID="cri-o://5f8b355b7009ffa37cb32fe8190cf0452a22787db1bc87dec02ef3d70ca2f80f" gracePeriod=30 Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.270741 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="06c6c196-5109-4edb-a034-bd5d45cb5873" containerName="nova-api-log" containerID="cri-o://597bdf8f502be320b1f3325fad9cc3ddffe4bfc96b395d70daec911a1c87e9cb" gracePeriod=30 Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.287157 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.315594 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" podStartSLOduration=3.315575996 podStartE2EDuration="3.315575996s" podCreationTimestamp="2025-10-03 16:46:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:46:31.313064762 +0000 UTC m=+1317.592940658" watchObservedRunningTime="2025-10-03 16:46:31.315575996 +0000 UTC m=+1317.595451882" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.426069 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-combined-ca-bundle\") pod \"2528f1c7-4a26-472c-84a5-c5d0a90c5ece\" (UID: \"2528f1c7-4a26-472c-84a5-c5d0a90c5ece\") " Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.426147 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea42249b-a5a1-4728-8be6-8babf3c51aa2-logs\") pod \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.426174 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkw7d\" (UniqueName: \"kubernetes.io/projected/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-kube-api-access-zkw7d\") pod \"2528f1c7-4a26-472c-84a5-c5d0a90c5ece\" (UID: \"2528f1c7-4a26-472c-84a5-c5d0a90c5ece\") " Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.426224 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-config-data\") pod \"2528f1c7-4a26-472c-84a5-c5d0a90c5ece\" (UID: \"2528f1c7-4a26-472c-84a5-c5d0a90c5ece\") " Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.426351 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea42249b-a5a1-4728-8be6-8babf3c51aa2-combined-ca-bundle\") pod \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.426403 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6x27\" (UniqueName: \"kubernetes.io/projected/ea42249b-a5a1-4728-8be6-8babf3c51aa2-kube-api-access-q6x27\") pod \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.426464 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea42249b-a5a1-4728-8be6-8babf3c51aa2-config-data\") pod \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\" (UID: \"ea42249b-a5a1-4728-8be6-8babf3c51aa2\") " Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.427093 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea42249b-a5a1-4728-8be6-8babf3c51aa2-logs" (OuterVolumeSpecName: "logs") pod "ea42249b-a5a1-4728-8be6-8babf3c51aa2" (UID: "ea42249b-a5a1-4728-8be6-8babf3c51aa2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.432201 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea42249b-a5a1-4728-8be6-8babf3c51aa2-kube-api-access-q6x27" (OuterVolumeSpecName: "kube-api-access-q6x27") pod "ea42249b-a5a1-4728-8be6-8babf3c51aa2" (UID: "ea42249b-a5a1-4728-8be6-8babf3c51aa2"). InnerVolumeSpecName "kube-api-access-q6x27". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.432224 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-kube-api-access-zkw7d" (OuterVolumeSpecName: "kube-api-access-zkw7d") pod "2528f1c7-4a26-472c-84a5-c5d0a90c5ece" (UID: "2528f1c7-4a26-472c-84a5-c5d0a90c5ece"). InnerVolumeSpecName "kube-api-access-zkw7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.456810 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea42249b-a5a1-4728-8be6-8babf3c51aa2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea42249b-a5a1-4728-8be6-8babf3c51aa2" (UID: "ea42249b-a5a1-4728-8be6-8babf3c51aa2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.457773 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea42249b-a5a1-4728-8be6-8babf3c51aa2-config-data" (OuterVolumeSpecName: "config-data") pod "ea42249b-a5a1-4728-8be6-8babf3c51aa2" (UID: "ea42249b-a5a1-4728-8be6-8babf3c51aa2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.458212 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-config-data" (OuterVolumeSpecName: "config-data") pod "2528f1c7-4a26-472c-84a5-c5d0a90c5ece" (UID: "2528f1c7-4a26-472c-84a5-c5d0a90c5ece"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.458848 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2528f1c7-4a26-472c-84a5-c5d0a90c5ece" (UID: "2528f1c7-4a26-472c-84a5-c5d0a90c5ece"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.529217 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea42249b-a5a1-4728-8be6-8babf3c51aa2-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.529256 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.529271 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea42249b-a5a1-4728-8be6-8babf3c51aa2-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.529280 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkw7d\" (UniqueName: \"kubernetes.io/projected/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-kube-api-access-zkw7d\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.529288 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2528f1c7-4a26-472c-84a5-c5d0a90c5ece-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.529296 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea42249b-a5a1-4728-8be6-8babf3c51aa2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:31 crc kubenswrapper[4744]: I1003 16:46:31.529304 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6x27\" (UniqueName: \"kubernetes.io/projected/ea42249b-a5a1-4728-8be6-8babf3c51aa2-kube-api-access-q6x27\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.281409 4744 generic.go:334] "Generic (PLEG): container finished" podID="06c6c196-5109-4edb-a034-bd5d45cb5873" containerID="597bdf8f502be320b1f3325fad9cc3ddffe4bfc96b395d70daec911a1c87e9cb" exitCode=143 Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.281531 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"06c6c196-5109-4edb-a034-bd5d45cb5873","Type":"ContainerDied","Data":"597bdf8f502be320b1f3325fad9cc3ddffe4bfc96b395d70daec911a1c87e9cb"} Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.281872 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.282005 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.340435 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.366244 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.411299 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.420557 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.428564 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 16:46:32 crc kubenswrapper[4744]: E1003 16:46:32.429108 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2528f1c7-4a26-472c-84a5-c5d0a90c5ece" containerName="nova-cell1-novncproxy-novncproxy" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.429129 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2528f1c7-4a26-472c-84a5-c5d0a90c5ece" containerName="nova-cell1-novncproxy-novncproxy" Oct 03 16:46:32 crc kubenswrapper[4744]: E1003 16:46:32.429144 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea42249b-a5a1-4728-8be6-8babf3c51aa2" containerName="nova-metadata-log" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.429154 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea42249b-a5a1-4728-8be6-8babf3c51aa2" containerName="nova-metadata-log" Oct 03 16:46:32 crc kubenswrapper[4744]: E1003 16:46:32.429167 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea42249b-a5a1-4728-8be6-8babf3c51aa2" containerName="nova-metadata-metadata" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.429174 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea42249b-a5a1-4728-8be6-8babf3c51aa2" containerName="nova-metadata-metadata" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.429442 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea42249b-a5a1-4728-8be6-8babf3c51aa2" containerName="nova-metadata-log" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.429460 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea42249b-a5a1-4728-8be6-8babf3c51aa2" containerName="nova-metadata-metadata" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.429478 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2528f1c7-4a26-472c-84a5-c5d0a90c5ece" containerName="nova-cell1-novncproxy-novncproxy" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.430275 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.433105 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.433342 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.438607 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.447864 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.449388 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.453696 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.454038 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.470183 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.481343 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.548128 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8005560a-4d21-4e19-9e47-17a703457765-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8005560a-4d21-4e19-9e47-17a703457765\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.548189 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.548265 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.548281 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f156453-e8c4-4106-bbfa-c25001c2f8e3-logs\") pod \"nova-metadata-0\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.548312 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-config-data\") pod \"nova-metadata-0\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.548346 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8005560a-4d21-4e19-9e47-17a703457765-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8005560a-4d21-4e19-9e47-17a703457765\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.548373 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8rqk\" (UniqueName: \"kubernetes.io/projected/4f156453-e8c4-4106-bbfa-c25001c2f8e3-kube-api-access-d8rqk\") pod \"nova-metadata-0\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.548414 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8005560a-4d21-4e19-9e47-17a703457765-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8005560a-4d21-4e19-9e47-17a703457765\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.548440 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8005560a-4d21-4e19-9e47-17a703457765-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8005560a-4d21-4e19-9e47-17a703457765\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.548462 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff8l7\" (UniqueName: \"kubernetes.io/projected/8005560a-4d21-4e19-9e47-17a703457765-kube-api-access-ff8l7\") pod \"nova-cell1-novncproxy-0\" (UID: \"8005560a-4d21-4e19-9e47-17a703457765\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.650594 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8005560a-4d21-4e19-9e47-17a703457765-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8005560a-4d21-4e19-9e47-17a703457765\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.650659 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.650723 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.650745 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f156453-e8c4-4106-bbfa-c25001c2f8e3-logs\") pod \"nova-metadata-0\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.650778 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-config-data\") pod \"nova-metadata-0\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.650810 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8005560a-4d21-4e19-9e47-17a703457765-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8005560a-4d21-4e19-9e47-17a703457765\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.650835 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8rqk\" (UniqueName: \"kubernetes.io/projected/4f156453-e8c4-4106-bbfa-c25001c2f8e3-kube-api-access-d8rqk\") pod \"nova-metadata-0\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.650882 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8005560a-4d21-4e19-9e47-17a703457765-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8005560a-4d21-4e19-9e47-17a703457765\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.650911 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8005560a-4d21-4e19-9e47-17a703457765-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8005560a-4d21-4e19-9e47-17a703457765\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.650938 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff8l7\" (UniqueName: \"kubernetes.io/projected/8005560a-4d21-4e19-9e47-17a703457765-kube-api-access-ff8l7\") pod \"nova-cell1-novncproxy-0\" (UID: \"8005560a-4d21-4e19-9e47-17a703457765\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.651555 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f156453-e8c4-4106-bbfa-c25001c2f8e3-logs\") pod \"nova-metadata-0\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.655641 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8005560a-4d21-4e19-9e47-17a703457765-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8005560a-4d21-4e19-9e47-17a703457765\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.656283 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.656473 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-config-data\") pod \"nova-metadata-0\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.657154 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.657732 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8005560a-4d21-4e19-9e47-17a703457765-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8005560a-4d21-4e19-9e47-17a703457765\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.658366 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8005560a-4d21-4e19-9e47-17a703457765-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8005560a-4d21-4e19-9e47-17a703457765\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.658913 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8005560a-4d21-4e19-9e47-17a703457765-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8005560a-4d21-4e19-9e47-17a703457765\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.670073 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8rqk\" (UniqueName: \"kubernetes.io/projected/4f156453-e8c4-4106-bbfa-c25001c2f8e3-kube-api-access-d8rqk\") pod \"nova-metadata-0\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " pod="openstack/nova-metadata-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.671331 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff8l7\" (UniqueName: \"kubernetes.io/projected/8005560a-4d21-4e19-9e47-17a703457765-kube-api-access-ff8l7\") pod \"nova-cell1-novncproxy-0\" (UID: \"8005560a-4d21-4e19-9e47-17a703457765\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.755938 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:32 crc kubenswrapper[4744]: I1003 16:46:32.781712 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 16:46:33 crc kubenswrapper[4744]: I1003 16:46:32.923692 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2528f1c7-4a26-472c-84a5-c5d0a90c5ece" path="/var/lib/kubelet/pods/2528f1c7-4a26-472c-84a5-c5d0a90c5ece/volumes" Oct 03 16:46:33 crc kubenswrapper[4744]: I1003 16:46:32.924960 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea42249b-a5a1-4728-8be6-8babf3c51aa2" path="/var/lib/kubelet/pods/ea42249b-a5a1-4728-8be6-8babf3c51aa2/volumes" Oct 03 16:46:33 crc kubenswrapper[4744]: I1003 16:46:33.836252 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 16:46:33 crc kubenswrapper[4744]: I1003 16:46:33.844613 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:46:33 crc kubenswrapper[4744]: W1003 16:46:33.851613 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f156453_e8c4_4106_bbfa_c25001c2f8e3.slice/crio-bda7a1ef5c834420b06ed75c81d3167e77cd67ef9c410dc7a8b08963797445c1 WatchSource:0}: Error finding container bda7a1ef5c834420b06ed75c81d3167e77cd67ef9c410dc7a8b08963797445c1: Status 404 returned error can't find the container with id bda7a1ef5c834420b06ed75c81d3167e77cd67ef9c410dc7a8b08963797445c1 Oct 03 16:46:33 crc kubenswrapper[4744]: W1003 16:46:33.856554 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8005560a_4d21_4e19_9e47_17a703457765.slice/crio-2297b81cacdda1d13bc35681927aa96310f55cbc07ba67a980a6e3907a4372d9 WatchSource:0}: Error finding container 2297b81cacdda1d13bc35681927aa96310f55cbc07ba67a980a6e3907a4372d9: Status 404 returned error can't find the container with id 2297b81cacdda1d13bc35681927aa96310f55cbc07ba67a980a6e3907a4372d9 Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.107195 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.283112 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-log-httpd\") pod \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.283208 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-combined-ca-bundle\") pod \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.283236 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-config-data\") pod \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.283285 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-ceilometer-tls-certs\") pod \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.283353 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-run-httpd\") pod \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.283422 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-sg-core-conf-yaml\") pod \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.283448 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-scripts\") pod \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.283489 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xq4d\" (UniqueName: \"kubernetes.io/projected/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-kube-api-access-6xq4d\") pod \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\" (UID: \"9d39052e-6ed5-47a7-9e91-2b47f55d94b0\") " Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.284159 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9d39052e-6ed5-47a7-9e91-2b47f55d94b0" (UID: "9d39052e-6ed5-47a7-9e91-2b47f55d94b0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.284375 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9d39052e-6ed5-47a7-9e91-2b47f55d94b0" (UID: "9d39052e-6ed5-47a7-9e91-2b47f55d94b0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.289230 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-kube-api-access-6xq4d" (OuterVolumeSpecName: "kube-api-access-6xq4d") pod "9d39052e-6ed5-47a7-9e91-2b47f55d94b0" (UID: "9d39052e-6ed5-47a7-9e91-2b47f55d94b0"). InnerVolumeSpecName "kube-api-access-6xq4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.293942 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-scripts" (OuterVolumeSpecName: "scripts") pod "9d39052e-6ed5-47a7-9e91-2b47f55d94b0" (UID: "9d39052e-6ed5-47a7-9e91-2b47f55d94b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.308913 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8005560a-4d21-4e19-9e47-17a703457765","Type":"ContainerStarted","Data":"d76fdd29d04446fb6bdb3bad24fb669c14070bc8f3417b4f9ff73bb09f5154f0"} Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.308965 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8005560a-4d21-4e19-9e47-17a703457765","Type":"ContainerStarted","Data":"2297b81cacdda1d13bc35681927aa96310f55cbc07ba67a980a6e3907a4372d9"} Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.313686 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4f156453-e8c4-4106-bbfa-c25001c2f8e3","Type":"ContainerStarted","Data":"943f82021ea5f083a16ea17129459a79e9148fc9ee81cc68bd5d4562e035c094"} Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.313726 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4f156453-e8c4-4106-bbfa-c25001c2f8e3","Type":"ContainerStarted","Data":"bda7a1ef5c834420b06ed75c81d3167e77cd67ef9c410dc7a8b08963797445c1"} Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.317529 4744 generic.go:334] "Generic (PLEG): container finished" podID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerID="6773f466de0370c325a944eb30b9da846e0dbbdd3cd8264d7000bda0b6b1680a" exitCode=0 Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.317559 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d39052e-6ed5-47a7-9e91-2b47f55d94b0","Type":"ContainerDied","Data":"6773f466de0370c325a944eb30b9da846e0dbbdd3cd8264d7000bda0b6b1680a"} Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.317576 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d39052e-6ed5-47a7-9e91-2b47f55d94b0","Type":"ContainerDied","Data":"a4ae2fc86dd093fea2daf39eb908eaecf4a3b15710ec846627d68cefa5df47d6"} Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.317592 4744 scope.go:117] "RemoveContainer" containerID="22c7ed5bc212ba6d0f0fd30d281e9f733ddda1db86129a7f3690833e47169b41" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.317708 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.336408 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9d39052e-6ed5-47a7-9e91-2b47f55d94b0" (UID: "9d39052e-6ed5-47a7-9e91-2b47f55d94b0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.341507 4744 scope.go:117] "RemoveContainer" containerID="9cd2c08fb236128fe62c1bd88e4109a7f4ac2235f7e5995c66fe8d10d7b7a10e" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.365551 4744 scope.go:117] "RemoveContainer" containerID="6773f466de0370c325a944eb30b9da846e0dbbdd3cd8264d7000bda0b6b1680a" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.381299 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "9d39052e-6ed5-47a7-9e91-2b47f55d94b0" (UID: "9d39052e-6ed5-47a7-9e91-2b47f55d94b0"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.385813 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.385853 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.385868 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.385879 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.385890 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.385900 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xq4d\" (UniqueName: \"kubernetes.io/projected/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-kube-api-access-6xq4d\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.391326 4744 scope.go:117] "RemoveContainer" containerID="405b4de4b9ea0cdd5e0dc6d254303bc0cac67c370ae38300acbf1e24cbd4b002" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.397718 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d39052e-6ed5-47a7-9e91-2b47f55d94b0" (UID: "9d39052e-6ed5-47a7-9e91-2b47f55d94b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.421716 4744 scope.go:117] "RemoveContainer" containerID="22c7ed5bc212ba6d0f0fd30d281e9f733ddda1db86129a7f3690833e47169b41" Oct 03 16:46:34 crc kubenswrapper[4744]: E1003 16:46:34.422309 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22c7ed5bc212ba6d0f0fd30d281e9f733ddda1db86129a7f3690833e47169b41\": container with ID starting with 22c7ed5bc212ba6d0f0fd30d281e9f733ddda1db86129a7f3690833e47169b41 not found: ID does not exist" containerID="22c7ed5bc212ba6d0f0fd30d281e9f733ddda1db86129a7f3690833e47169b41" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.422423 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22c7ed5bc212ba6d0f0fd30d281e9f733ddda1db86129a7f3690833e47169b41"} err="failed to get container status \"22c7ed5bc212ba6d0f0fd30d281e9f733ddda1db86129a7f3690833e47169b41\": rpc error: code = NotFound desc = could not find container \"22c7ed5bc212ba6d0f0fd30d281e9f733ddda1db86129a7f3690833e47169b41\": container with ID starting with 22c7ed5bc212ba6d0f0fd30d281e9f733ddda1db86129a7f3690833e47169b41 not found: ID does not exist" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.422551 4744 scope.go:117] "RemoveContainer" containerID="9cd2c08fb236128fe62c1bd88e4109a7f4ac2235f7e5995c66fe8d10d7b7a10e" Oct 03 16:46:34 crc kubenswrapper[4744]: E1003 16:46:34.424399 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cd2c08fb236128fe62c1bd88e4109a7f4ac2235f7e5995c66fe8d10d7b7a10e\": container with ID starting with 9cd2c08fb236128fe62c1bd88e4109a7f4ac2235f7e5995c66fe8d10d7b7a10e not found: ID does not exist" containerID="9cd2c08fb236128fe62c1bd88e4109a7f4ac2235f7e5995c66fe8d10d7b7a10e" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.424457 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cd2c08fb236128fe62c1bd88e4109a7f4ac2235f7e5995c66fe8d10d7b7a10e"} err="failed to get container status \"9cd2c08fb236128fe62c1bd88e4109a7f4ac2235f7e5995c66fe8d10d7b7a10e\": rpc error: code = NotFound desc = could not find container \"9cd2c08fb236128fe62c1bd88e4109a7f4ac2235f7e5995c66fe8d10d7b7a10e\": container with ID starting with 9cd2c08fb236128fe62c1bd88e4109a7f4ac2235f7e5995c66fe8d10d7b7a10e not found: ID does not exist" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.424503 4744 scope.go:117] "RemoveContainer" containerID="6773f466de0370c325a944eb30b9da846e0dbbdd3cd8264d7000bda0b6b1680a" Oct 03 16:46:34 crc kubenswrapper[4744]: E1003 16:46:34.428901 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6773f466de0370c325a944eb30b9da846e0dbbdd3cd8264d7000bda0b6b1680a\": container with ID starting with 6773f466de0370c325a944eb30b9da846e0dbbdd3cd8264d7000bda0b6b1680a not found: ID does not exist" containerID="6773f466de0370c325a944eb30b9da846e0dbbdd3cd8264d7000bda0b6b1680a" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.429084 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6773f466de0370c325a944eb30b9da846e0dbbdd3cd8264d7000bda0b6b1680a"} err="failed to get container status \"6773f466de0370c325a944eb30b9da846e0dbbdd3cd8264d7000bda0b6b1680a\": rpc error: code = NotFound desc = could not find container \"6773f466de0370c325a944eb30b9da846e0dbbdd3cd8264d7000bda0b6b1680a\": container with ID starting with 6773f466de0370c325a944eb30b9da846e0dbbdd3cd8264d7000bda0b6b1680a not found: ID does not exist" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.429181 4744 scope.go:117] "RemoveContainer" containerID="405b4de4b9ea0cdd5e0dc6d254303bc0cac67c370ae38300acbf1e24cbd4b002" Oct 03 16:46:34 crc kubenswrapper[4744]: E1003 16:46:34.430336 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"405b4de4b9ea0cdd5e0dc6d254303bc0cac67c370ae38300acbf1e24cbd4b002\": container with ID starting with 405b4de4b9ea0cdd5e0dc6d254303bc0cac67c370ae38300acbf1e24cbd4b002 not found: ID does not exist" containerID="405b4de4b9ea0cdd5e0dc6d254303bc0cac67c370ae38300acbf1e24cbd4b002" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.430411 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"405b4de4b9ea0cdd5e0dc6d254303bc0cac67c370ae38300acbf1e24cbd4b002"} err="failed to get container status \"405b4de4b9ea0cdd5e0dc6d254303bc0cac67c370ae38300acbf1e24cbd4b002\": rpc error: code = NotFound desc = could not find container \"405b4de4b9ea0cdd5e0dc6d254303bc0cac67c370ae38300acbf1e24cbd4b002\": container with ID starting with 405b4de4b9ea0cdd5e0dc6d254303bc0cac67c370ae38300acbf1e24cbd4b002 not found: ID does not exist" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.448169 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-config-data" (OuterVolumeSpecName: "config-data") pod "9d39052e-6ed5-47a7-9e91-2b47f55d94b0" (UID: "9d39052e-6ed5-47a7-9e91-2b47f55d94b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.487592 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.487644 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d39052e-6ed5-47a7-9e91-2b47f55d94b0-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.653062 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.653039032 podStartE2EDuration="2.653039032s" podCreationTimestamp="2025-10-03 16:46:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:46:34.3308151 +0000 UTC m=+1320.610690996" watchObservedRunningTime="2025-10-03 16:46:34.653039032 +0000 UTC m=+1320.932914928" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.660584 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.669209 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.703036 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:46:34 crc kubenswrapper[4744]: E1003 16:46:34.703444 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="ceilometer-central-agent" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.703462 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="ceilometer-central-agent" Oct 03 16:46:34 crc kubenswrapper[4744]: E1003 16:46:34.703508 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="proxy-httpd" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.703515 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="proxy-httpd" Oct 03 16:46:34 crc kubenswrapper[4744]: E1003 16:46:34.703535 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="ceilometer-notification-agent" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.703541 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="ceilometer-notification-agent" Oct 03 16:46:34 crc kubenswrapper[4744]: E1003 16:46:34.703549 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="sg-core" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.703555 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="sg-core" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.703725 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="ceilometer-notification-agent" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.703751 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="ceilometer-central-agent" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.703773 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="proxy-httpd" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.703788 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" containerName="sg-core" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.705800 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.710850 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.711300 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.711420 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.743658 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.786175 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.801868 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13a29aaf-7761-4250-b9fd-bb69661f450e-scripts\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.801928 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13a29aaf-7761-4250-b9fd-bb69661f450e-config-data\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.801956 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13a29aaf-7761-4250-b9fd-bb69661f450e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.802008 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13a29aaf-7761-4250-b9fd-bb69661f450e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.802055 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq42j\" (UniqueName: \"kubernetes.io/projected/13a29aaf-7761-4250-b9fd-bb69661f450e-kube-api-access-qq42j\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.802081 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13a29aaf-7761-4250-b9fd-bb69661f450e-run-httpd\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.802112 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13a29aaf-7761-4250-b9fd-bb69661f450e-log-httpd\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.802262 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/13a29aaf-7761-4250-b9fd-bb69661f450e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.903327 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06c6c196-5109-4edb-a034-bd5d45cb5873-config-data\") pod \"06c6c196-5109-4edb-a034-bd5d45cb5873\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.903608 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qg2q\" (UniqueName: \"kubernetes.io/projected/06c6c196-5109-4edb-a034-bd5d45cb5873-kube-api-access-5qg2q\") pod \"06c6c196-5109-4edb-a034-bd5d45cb5873\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.903702 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06c6c196-5109-4edb-a034-bd5d45cb5873-logs\") pod \"06c6c196-5109-4edb-a034-bd5d45cb5873\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.903753 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c6c196-5109-4edb-a034-bd5d45cb5873-combined-ca-bundle\") pod \"06c6c196-5109-4edb-a034-bd5d45cb5873\" (UID: \"06c6c196-5109-4edb-a034-bd5d45cb5873\") " Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.903961 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13a29aaf-7761-4250-b9fd-bb69661f450e-run-httpd\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.903997 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13a29aaf-7761-4250-b9fd-bb69661f450e-log-httpd\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.905270 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/13a29aaf-7761-4250-b9fd-bb69661f450e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.905350 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13a29aaf-7761-4250-b9fd-bb69661f450e-scripts\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.905382 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13a29aaf-7761-4250-b9fd-bb69661f450e-config-data\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.905404 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13a29aaf-7761-4250-b9fd-bb69661f450e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.905888 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13a29aaf-7761-4250-b9fd-bb69661f450e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.907952 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq42j\" (UniqueName: \"kubernetes.io/projected/13a29aaf-7761-4250-b9fd-bb69661f450e-kube-api-access-qq42j\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.907426 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.905940 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13a29aaf-7761-4250-b9fd-bb69661f450e-log-httpd\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.907555 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.909864 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.913565 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06c6c196-5109-4edb-a034-bd5d45cb5873-logs" (OuterVolumeSpecName: "logs") pod "06c6c196-5109-4edb-a034-bd5d45cb5873" (UID: "06c6c196-5109-4edb-a034-bd5d45cb5873"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.915158 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06c6c196-5109-4edb-a034-bd5d45cb5873-kube-api-access-5qg2q" (OuterVolumeSpecName: "kube-api-access-5qg2q") pod "06c6c196-5109-4edb-a034-bd5d45cb5873" (UID: "06c6c196-5109-4edb-a034-bd5d45cb5873"). InnerVolumeSpecName "kube-api-access-5qg2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.925152 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/13a29aaf-7761-4250-b9fd-bb69661f450e-run-httpd\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.928959 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13a29aaf-7761-4250-b9fd-bb69661f450e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.933958 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d39052e-6ed5-47a7-9e91-2b47f55d94b0" path="/var/lib/kubelet/pods/9d39052e-6ed5-47a7-9e91-2b47f55d94b0/volumes" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.937131 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/13a29aaf-7761-4250-b9fd-bb69661f450e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.939461 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq42j\" (UniqueName: \"kubernetes.io/projected/13a29aaf-7761-4250-b9fd-bb69661f450e-kube-api-access-qq42j\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.944803 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/13a29aaf-7761-4250-b9fd-bb69661f450e-scripts\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.946898 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13a29aaf-7761-4250-b9fd-bb69661f450e-config-data\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.963411 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/13a29aaf-7761-4250-b9fd-bb69661f450e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"13a29aaf-7761-4250-b9fd-bb69661f450e\") " pod="openstack/ceilometer-0" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.964398 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06c6c196-5109-4edb-a034-bd5d45cb5873-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06c6c196-5109-4edb-a034-bd5d45cb5873" (UID: "06c6c196-5109-4edb-a034-bd5d45cb5873"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:34 crc kubenswrapper[4744]: I1003 16:46:34.983155 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06c6c196-5109-4edb-a034-bd5d45cb5873-config-data" (OuterVolumeSpecName: "config-data") pod "06c6c196-5109-4edb-a034-bd5d45cb5873" (UID: "06c6c196-5109-4edb-a034-bd5d45cb5873"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.011114 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qg2q\" (UniqueName: \"kubernetes.io/projected/06c6c196-5109-4edb-a034-bd5d45cb5873-kube-api-access-5qg2q\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.011178 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06c6c196-5109-4edb-a034-bd5d45cb5873-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.011193 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06c6c196-5109-4edb-a034-bd5d45cb5873-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.011204 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06c6c196-5109-4edb-a034-bd5d45cb5873-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.023473 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.328031 4744 generic.go:334] "Generic (PLEG): container finished" podID="06c6c196-5109-4edb-a034-bd5d45cb5873" containerID="5f8b355b7009ffa37cb32fe8190cf0452a22787db1bc87dec02ef3d70ca2f80f" exitCode=0 Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.328087 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"06c6c196-5109-4edb-a034-bd5d45cb5873","Type":"ContainerDied","Data":"5f8b355b7009ffa37cb32fe8190cf0452a22787db1bc87dec02ef3d70ca2f80f"} Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.328115 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"06c6c196-5109-4edb-a034-bd5d45cb5873","Type":"ContainerDied","Data":"95c35046d05dff5b9b6440c91e4012edf4cfd208c48b46ce2d3d52a950be4c5b"} Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.328131 4744 scope.go:117] "RemoveContainer" containerID="5f8b355b7009ffa37cb32fe8190cf0452a22787db1bc87dec02ef3d70ca2f80f" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.328212 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.339115 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4f156453-e8c4-4106-bbfa-c25001c2f8e3","Type":"ContainerStarted","Data":"eca1c37f4ad609b0b529de2418979bcb6996fae3c95d43e07349a686bbf8bced"} Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.375256 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.375234852 podStartE2EDuration="3.375234852s" podCreationTimestamp="2025-10-03 16:46:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:46:35.364051058 +0000 UTC m=+1321.643926954" watchObservedRunningTime="2025-10-03 16:46:35.375234852 +0000 UTC m=+1321.655110748" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.376469 4744 scope.go:117] "RemoveContainer" containerID="597bdf8f502be320b1f3325fad9cc3ddffe4bfc96b395d70daec911a1c87e9cb" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.398231 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.407547 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.408293 4744 scope.go:117] "RemoveContainer" containerID="5f8b355b7009ffa37cb32fe8190cf0452a22787db1bc87dec02ef3d70ca2f80f" Oct 03 16:46:35 crc kubenswrapper[4744]: E1003 16:46:35.408937 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f8b355b7009ffa37cb32fe8190cf0452a22787db1bc87dec02ef3d70ca2f80f\": container with ID starting with 5f8b355b7009ffa37cb32fe8190cf0452a22787db1bc87dec02ef3d70ca2f80f not found: ID does not exist" containerID="5f8b355b7009ffa37cb32fe8190cf0452a22787db1bc87dec02ef3d70ca2f80f" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.409040 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f8b355b7009ffa37cb32fe8190cf0452a22787db1bc87dec02ef3d70ca2f80f"} err="failed to get container status \"5f8b355b7009ffa37cb32fe8190cf0452a22787db1bc87dec02ef3d70ca2f80f\": rpc error: code = NotFound desc = could not find container \"5f8b355b7009ffa37cb32fe8190cf0452a22787db1bc87dec02ef3d70ca2f80f\": container with ID starting with 5f8b355b7009ffa37cb32fe8190cf0452a22787db1bc87dec02ef3d70ca2f80f not found: ID does not exist" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.409253 4744 scope.go:117] "RemoveContainer" containerID="597bdf8f502be320b1f3325fad9cc3ddffe4bfc96b395d70daec911a1c87e9cb" Oct 03 16:46:35 crc kubenswrapper[4744]: E1003 16:46:35.409719 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"597bdf8f502be320b1f3325fad9cc3ddffe4bfc96b395d70daec911a1c87e9cb\": container with ID starting with 597bdf8f502be320b1f3325fad9cc3ddffe4bfc96b395d70daec911a1c87e9cb not found: ID does not exist" containerID="597bdf8f502be320b1f3325fad9cc3ddffe4bfc96b395d70daec911a1c87e9cb" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.409972 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"597bdf8f502be320b1f3325fad9cc3ddffe4bfc96b395d70daec911a1c87e9cb"} err="failed to get container status \"597bdf8f502be320b1f3325fad9cc3ddffe4bfc96b395d70daec911a1c87e9cb\": rpc error: code = NotFound desc = could not find container \"597bdf8f502be320b1f3325fad9cc3ddffe4bfc96b395d70daec911a1c87e9cb\": container with ID starting with 597bdf8f502be320b1f3325fad9cc3ddffe4bfc96b395d70daec911a1c87e9cb not found: ID does not exist" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.420630 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:35 crc kubenswrapper[4744]: E1003 16:46:35.421177 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06c6c196-5109-4edb-a034-bd5d45cb5873" containerName="nova-api-api" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.421199 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="06c6c196-5109-4edb-a034-bd5d45cb5873" containerName="nova-api-api" Oct 03 16:46:35 crc kubenswrapper[4744]: E1003 16:46:35.421249 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06c6c196-5109-4edb-a034-bd5d45cb5873" containerName="nova-api-log" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.421258 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="06c6c196-5109-4edb-a034-bd5d45cb5873" containerName="nova-api-log" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.421573 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="06c6c196-5109-4edb-a034-bd5d45cb5873" containerName="nova-api-api" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.421607 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="06c6c196-5109-4edb-a034-bd5d45cb5873" containerName="nova-api-log" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.423265 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.437351 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.437797 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.437953 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.448911 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:35 crc kubenswrapper[4744]: W1003 16:46:35.495732 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13a29aaf_7761_4250_b9fd_bb69661f450e.slice/crio-470dcd0e89f423fb315f31787271c315ab5c3ea8e3869d983f396e0864a45114 WatchSource:0}: Error finding container 470dcd0e89f423fb315f31787271c315ab5c3ea8e3869d983f396e0864a45114: Status 404 returned error can't find the container with id 470dcd0e89f423fb315f31787271c315ab5c3ea8e3869d983f396e0864a45114 Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.498369 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.521894 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1e20b1a-3009-4bcb-b936-39aade55a6d0-logs\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.521958 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-config-data\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.522441 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-public-tls-certs\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.522640 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cq8r\" (UniqueName: \"kubernetes.io/projected/c1e20b1a-3009-4bcb-b936-39aade55a6d0-kube-api-access-9cq8r\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.522693 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.522788 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.624143 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-public-tls-certs\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.624242 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cq8r\" (UniqueName: \"kubernetes.io/projected/c1e20b1a-3009-4bcb-b936-39aade55a6d0-kube-api-access-9cq8r\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.624276 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.624320 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.624347 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1e20b1a-3009-4bcb-b936-39aade55a6d0-logs\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.624391 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-config-data\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.625504 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1e20b1a-3009-4bcb-b936-39aade55a6d0-logs\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.629737 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.631381 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-public-tls-certs\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.631443 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-config-data\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.631817 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.644642 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cq8r\" (UniqueName: \"kubernetes.io/projected/c1e20b1a-3009-4bcb-b936-39aade55a6d0-kube-api-access-9cq8r\") pod \"nova-api-0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " pod="openstack/nova-api-0" Oct 03 16:46:35 crc kubenswrapper[4744]: I1003 16:46:35.752290 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:46:36 crc kubenswrapper[4744]: I1003 16:46:36.243656 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:36 crc kubenswrapper[4744]: I1003 16:46:36.348880 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1e20b1a-3009-4bcb-b936-39aade55a6d0","Type":"ContainerStarted","Data":"c27c0ffbee0a6cac85b3f5454657989d60c374a431a1c308969015c18282b50a"} Oct 03 16:46:36 crc kubenswrapper[4744]: I1003 16:46:36.350876 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13a29aaf-7761-4250-b9fd-bb69661f450e","Type":"ContainerStarted","Data":"6b639afda00ba7f68b0169b18f55d792bd84517caccd8e18977e8c2b9de8f4b2"} Oct 03 16:46:36 crc kubenswrapper[4744]: I1003 16:46:36.350921 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13a29aaf-7761-4250-b9fd-bb69661f450e","Type":"ContainerStarted","Data":"470dcd0e89f423fb315f31787271c315ab5c3ea8e3869d983f396e0864a45114"} Oct 03 16:46:36 crc kubenswrapper[4744]: I1003 16:46:36.902923 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06c6c196-5109-4edb-a034-bd5d45cb5873" path="/var/lib/kubelet/pods/06c6c196-5109-4edb-a034-bd5d45cb5873/volumes" Oct 03 16:46:37 crc kubenswrapper[4744]: I1003 16:46:37.362204 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1e20b1a-3009-4bcb-b936-39aade55a6d0","Type":"ContainerStarted","Data":"f0fa717771b2744e99974d5a0a38277c3759cd0f335c2e7074450c71721a16ba"} Oct 03 16:46:37 crc kubenswrapper[4744]: I1003 16:46:37.362440 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1e20b1a-3009-4bcb-b936-39aade55a6d0","Type":"ContainerStarted","Data":"0407bd8c6d71b297d98817f30c443032c1c6bf771addc7a789a8ec0cc81c0e98"} Oct 03 16:46:37 crc kubenswrapper[4744]: I1003 16:46:37.366146 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13a29aaf-7761-4250-b9fd-bb69661f450e","Type":"ContainerStarted","Data":"edf89dbd9039bd3b01316ca92dcf52a8a229763a6222d371cf0869d29b908609"} Oct 03 16:46:37 crc kubenswrapper[4744]: I1003 16:46:37.410248 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.410226766 podStartE2EDuration="2.410226766s" podCreationTimestamp="2025-10-03 16:46:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:46:37.381177958 +0000 UTC m=+1323.661053854" watchObservedRunningTime="2025-10-03 16:46:37.410226766 +0000 UTC m=+1323.690102662" Oct 03 16:46:37 crc kubenswrapper[4744]: I1003 16:46:37.756727 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:37 crc kubenswrapper[4744]: I1003 16:46:37.782683 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 16:46:37 crc kubenswrapper[4744]: I1003 16:46:37.782748 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 16:46:38 crc kubenswrapper[4744]: I1003 16:46:38.382727 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13a29aaf-7761-4250-b9fd-bb69661f450e","Type":"ContainerStarted","Data":"239a19898f927ba5272c67d7cf429836b65d2154f19dc8423739f0c6cf5116c4"} Oct 03 16:46:38 crc kubenswrapper[4744]: I1003 16:46:38.746687 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:46:38 crc kubenswrapper[4744]: I1003 16:46:38.837719 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs"] Oct 03 16:46:38 crc kubenswrapper[4744]: I1003 16:46:38.838284 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" podUID="29c874fa-abab-46d4-99e5-b44e6332f71d" containerName="dnsmasq-dns" containerID="cri-o://c84611e4217cd7d9a6a73bfcb2bfef411bb121c02dc022732dd5af8dbc5eef5d" gracePeriod=10 Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.340062 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.395219 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"13a29aaf-7761-4250-b9fd-bb69661f450e","Type":"ContainerStarted","Data":"9e021efd319acacbb5591323b2789afe8432fbd3715e27e9ffda2bb6165a6470"} Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.395453 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.396919 4744 generic.go:334] "Generic (PLEG): container finished" podID="29c874fa-abab-46d4-99e5-b44e6332f71d" containerID="c84611e4217cd7d9a6a73bfcb2bfef411bb121c02dc022732dd5af8dbc5eef5d" exitCode=0 Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.396963 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" event={"ID":"29c874fa-abab-46d4-99e5-b44e6332f71d","Type":"ContainerDied","Data":"c84611e4217cd7d9a6a73bfcb2bfef411bb121c02dc022732dd5af8dbc5eef5d"} Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.396994 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" event={"ID":"29c874fa-abab-46d4-99e5-b44e6332f71d","Type":"ContainerDied","Data":"a23ffcd21f98b404999bfdd5a18c5b44cff652def773b7402b44461350d1bee8"} Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.397014 4744 scope.go:117] "RemoveContainer" containerID="c84611e4217cd7d9a6a73bfcb2bfef411bb121c02dc022732dd5af8dbc5eef5d" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.397153 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.406210 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-ovsdbserver-nb\") pod \"29c874fa-abab-46d4-99e5-b44e6332f71d\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.406301 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-config\") pod \"29c874fa-abab-46d4-99e5-b44e6332f71d\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.406380 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-dns-swift-storage-0\") pod \"29c874fa-abab-46d4-99e5-b44e6332f71d\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.406423 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-dns-svc\") pod \"29c874fa-abab-46d4-99e5-b44e6332f71d\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.406536 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq4lz\" (UniqueName: \"kubernetes.io/projected/29c874fa-abab-46d4-99e5-b44e6332f71d-kube-api-access-mq4lz\") pod \"29c874fa-abab-46d4-99e5-b44e6332f71d\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.406677 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-ovsdbserver-sb\") pod \"29c874fa-abab-46d4-99e5-b44e6332f71d\" (UID: \"29c874fa-abab-46d4-99e5-b44e6332f71d\") " Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.419663 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29c874fa-abab-46d4-99e5-b44e6332f71d-kube-api-access-mq4lz" (OuterVolumeSpecName: "kube-api-access-mq4lz") pod "29c874fa-abab-46d4-99e5-b44e6332f71d" (UID: "29c874fa-abab-46d4-99e5-b44e6332f71d"). InnerVolumeSpecName "kube-api-access-mq4lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.463319 4744 scope.go:117] "RemoveContainer" containerID="54759bc2ac35a0adf4ad31b0b2b040444896aa440fd9a749dc4119f740569c27" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.470202 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-config" (OuterVolumeSpecName: "config") pod "29c874fa-abab-46d4-99e5-b44e6332f71d" (UID: "29c874fa-abab-46d4-99e5-b44e6332f71d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.481768 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "29c874fa-abab-46d4-99e5-b44e6332f71d" (UID: "29c874fa-abab-46d4-99e5-b44e6332f71d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.487116 4744 scope.go:117] "RemoveContainer" containerID="c84611e4217cd7d9a6a73bfcb2bfef411bb121c02dc022732dd5af8dbc5eef5d" Oct 03 16:46:39 crc kubenswrapper[4744]: E1003 16:46:39.487560 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c84611e4217cd7d9a6a73bfcb2bfef411bb121c02dc022732dd5af8dbc5eef5d\": container with ID starting with c84611e4217cd7d9a6a73bfcb2bfef411bb121c02dc022732dd5af8dbc5eef5d not found: ID does not exist" containerID="c84611e4217cd7d9a6a73bfcb2bfef411bb121c02dc022732dd5af8dbc5eef5d" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.487596 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c84611e4217cd7d9a6a73bfcb2bfef411bb121c02dc022732dd5af8dbc5eef5d"} err="failed to get container status \"c84611e4217cd7d9a6a73bfcb2bfef411bb121c02dc022732dd5af8dbc5eef5d\": rpc error: code = NotFound desc = could not find container \"c84611e4217cd7d9a6a73bfcb2bfef411bb121c02dc022732dd5af8dbc5eef5d\": container with ID starting with c84611e4217cd7d9a6a73bfcb2bfef411bb121c02dc022732dd5af8dbc5eef5d not found: ID does not exist" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.487621 4744 scope.go:117] "RemoveContainer" containerID="54759bc2ac35a0adf4ad31b0b2b040444896aa440fd9a749dc4119f740569c27" Oct 03 16:46:39 crc kubenswrapper[4744]: E1003 16:46:39.487863 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54759bc2ac35a0adf4ad31b0b2b040444896aa440fd9a749dc4119f740569c27\": container with ID starting with 54759bc2ac35a0adf4ad31b0b2b040444896aa440fd9a749dc4119f740569c27 not found: ID does not exist" containerID="54759bc2ac35a0adf4ad31b0b2b040444896aa440fd9a749dc4119f740569c27" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.487892 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54759bc2ac35a0adf4ad31b0b2b040444896aa440fd9a749dc4119f740569c27"} err="failed to get container status \"54759bc2ac35a0adf4ad31b0b2b040444896aa440fd9a749dc4119f740569c27\": rpc error: code = NotFound desc = could not find container \"54759bc2ac35a0adf4ad31b0b2b040444896aa440fd9a749dc4119f740569c27\": container with ID starting with 54759bc2ac35a0adf4ad31b0b2b040444896aa440fd9a749dc4119f740569c27 not found: ID does not exist" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.492122 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "29c874fa-abab-46d4-99e5-b44e6332f71d" (UID: "29c874fa-abab-46d4-99e5-b44e6332f71d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.495675 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "29c874fa-abab-46d4-99e5-b44e6332f71d" (UID: "29c874fa-abab-46d4-99e5-b44e6332f71d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.509243 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.509270 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.509279 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.509289 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.509299 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq4lz\" (UniqueName: \"kubernetes.io/projected/29c874fa-abab-46d4-99e5-b44e6332f71d-kube-api-access-mq4lz\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.510851 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "29c874fa-abab-46d4-99e5-b44e6332f71d" (UID: "29c874fa-abab-46d4-99e5-b44e6332f71d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.612760 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/29c874fa-abab-46d4-99e5-b44e6332f71d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.736115 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.236867627 podStartE2EDuration="5.736088865s" podCreationTimestamp="2025-10-03 16:46:34 +0000 UTC" firstStartedPulling="2025-10-03 16:46:35.498142597 +0000 UTC m=+1321.778018493" lastFinishedPulling="2025-10-03 16:46:38.997363835 +0000 UTC m=+1325.277239731" observedRunningTime="2025-10-03 16:46:39.431876572 +0000 UTC m=+1325.711752468" watchObservedRunningTime="2025-10-03 16:46:39.736088865 +0000 UTC m=+1326.015964801" Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.741042 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs"] Oct 03 16:46:39 crc kubenswrapper[4744]: I1003 16:46:39.751139 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d5fbbb8c5-xjwgs"] Oct 03 16:46:40 crc kubenswrapper[4744]: I1003 16:46:40.915187 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29c874fa-abab-46d4-99e5-b44e6332f71d" path="/var/lib/kubelet/pods/29c874fa-abab-46d4-99e5-b44e6332f71d/volumes" Oct 03 16:46:42 crc kubenswrapper[4744]: I1003 16:46:42.756361 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:42 crc kubenswrapper[4744]: I1003 16:46:42.782799 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 16:46:42 crc kubenswrapper[4744]: I1003 16:46:42.782861 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 16:46:42 crc kubenswrapper[4744]: I1003 16:46:42.793972 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.457322 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.613961 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-h7gcm"] Oct 03 16:46:43 crc kubenswrapper[4744]: E1003 16:46:43.614781 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29c874fa-abab-46d4-99e5-b44e6332f71d" containerName="init" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.614802 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="29c874fa-abab-46d4-99e5-b44e6332f71d" containerName="init" Oct 03 16:46:43 crc kubenswrapper[4744]: E1003 16:46:43.614810 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29c874fa-abab-46d4-99e5-b44e6332f71d" containerName="dnsmasq-dns" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.614819 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="29c874fa-abab-46d4-99e5-b44e6332f71d" containerName="dnsmasq-dns" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.615012 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="29c874fa-abab-46d4-99e5-b44e6332f71d" containerName="dnsmasq-dns" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.615691 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.619154 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.619444 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.624891 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-h7gcm"] Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.702739 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-config-data\") pod \"nova-cell1-cell-mapping-h7gcm\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.702823 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkpgj\" (UniqueName: \"kubernetes.io/projected/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-kube-api-access-rkpgj\") pod \"nova-cell1-cell-mapping-h7gcm\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.702895 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-scripts\") pod \"nova-cell1-cell-mapping-h7gcm\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.702974 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-h7gcm\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.798653 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4f156453-e8c4-4106-bbfa-c25001c2f8e3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.212:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.798675 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4f156453-e8c4-4106-bbfa-c25001c2f8e3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.212:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.804594 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-scripts\") pod \"nova-cell1-cell-mapping-h7gcm\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.804792 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-h7gcm\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.804969 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-config-data\") pod \"nova-cell1-cell-mapping-h7gcm\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.805093 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkpgj\" (UniqueName: \"kubernetes.io/projected/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-kube-api-access-rkpgj\") pod \"nova-cell1-cell-mapping-h7gcm\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.818182 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-scripts\") pod \"nova-cell1-cell-mapping-h7gcm\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.818295 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-config-data\") pod \"nova-cell1-cell-mapping-h7gcm\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.818659 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-h7gcm\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.832187 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkpgj\" (UniqueName: \"kubernetes.io/projected/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-kube-api-access-rkpgj\") pod \"nova-cell1-cell-mapping-h7gcm\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:43 crc kubenswrapper[4744]: I1003 16:46:43.935800 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:44 crc kubenswrapper[4744]: I1003 16:46:44.443578 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-h7gcm"] Oct 03 16:46:45 crc kubenswrapper[4744]: I1003 16:46:45.462743 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-h7gcm" event={"ID":"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d","Type":"ContainerStarted","Data":"147f7ab564c52abc13651319d27f4b237440be77a2aa5fc85cec734680c0a2cf"} Oct 03 16:46:45 crc kubenswrapper[4744]: I1003 16:46:45.462797 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-h7gcm" event={"ID":"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d","Type":"ContainerStarted","Data":"96fb4e1f4da526608191036ab9d803165801b7be800a8dbe48a64c537846eb38"} Oct 03 16:46:45 crc kubenswrapper[4744]: I1003 16:46:45.485038 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-h7gcm" podStartSLOduration=2.485022467 podStartE2EDuration="2.485022467s" podCreationTimestamp="2025-10-03 16:46:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:46:45.484082083 +0000 UTC m=+1331.763957989" watchObservedRunningTime="2025-10-03 16:46:45.485022467 +0000 UTC m=+1331.764898363" Oct 03 16:46:45 crc kubenswrapper[4744]: I1003 16:46:45.753482 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 16:46:45 crc kubenswrapper[4744]: I1003 16:46:45.753632 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 16:46:46 crc kubenswrapper[4744]: I1003 16:46:46.771909 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c1e20b1a-3009-4bcb-b936-39aade55a6d0" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.214:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 16:46:46 crc kubenswrapper[4744]: I1003 16:46:46.772183 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c1e20b1a-3009-4bcb-b936-39aade55a6d0" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.214:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 16:46:49 crc kubenswrapper[4744]: I1003 16:46:49.517958 4744 generic.go:334] "Generic (PLEG): container finished" podID="e2cdd571-e0fc-40c3-bf1f-40bfcf91913d" containerID="147f7ab564c52abc13651319d27f4b237440be77a2aa5fc85cec734680c0a2cf" exitCode=0 Oct 03 16:46:49 crc kubenswrapper[4744]: I1003 16:46:49.518050 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-h7gcm" event={"ID":"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d","Type":"ContainerDied","Data":"147f7ab564c52abc13651319d27f4b237440be77a2aa5fc85cec734680c0a2cf"} Oct 03 16:46:50 crc kubenswrapper[4744]: I1003 16:46:50.998069 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.066524 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-config-data\") pod \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.066867 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-combined-ca-bundle\") pod \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.067018 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkpgj\" (UniqueName: \"kubernetes.io/projected/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-kube-api-access-rkpgj\") pod \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.067061 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-scripts\") pod \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\" (UID: \"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d\") " Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.074805 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-kube-api-access-rkpgj" (OuterVolumeSpecName: "kube-api-access-rkpgj") pod "e2cdd571-e0fc-40c3-bf1f-40bfcf91913d" (UID: "e2cdd571-e0fc-40c3-bf1f-40bfcf91913d"). InnerVolumeSpecName "kube-api-access-rkpgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.074887 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-scripts" (OuterVolumeSpecName: "scripts") pod "e2cdd571-e0fc-40c3-bf1f-40bfcf91913d" (UID: "e2cdd571-e0fc-40c3-bf1f-40bfcf91913d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.100088 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-config-data" (OuterVolumeSpecName: "config-data") pod "e2cdd571-e0fc-40c3-bf1f-40bfcf91913d" (UID: "e2cdd571-e0fc-40c3-bf1f-40bfcf91913d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.125693 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2cdd571-e0fc-40c3-bf1f-40bfcf91913d" (UID: "e2cdd571-e0fc-40c3-bf1f-40bfcf91913d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.169148 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.169185 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.169198 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkpgj\" (UniqueName: \"kubernetes.io/projected/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-kube-api-access-rkpgj\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.169207 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.544341 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-h7gcm" event={"ID":"e2cdd571-e0fc-40c3-bf1f-40bfcf91913d","Type":"ContainerDied","Data":"96fb4e1f4da526608191036ab9d803165801b7be800a8dbe48a64c537846eb38"} Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.544376 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96fb4e1f4da526608191036ab9d803165801b7be800a8dbe48a64c537846eb38" Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.544390 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-h7gcm" Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.758290 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.758631 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c1e20b1a-3009-4bcb-b936-39aade55a6d0" containerName="nova-api-log" containerID="cri-o://0407bd8c6d71b297d98817f30c443032c1c6bf771addc7a789a8ec0cc81c0e98" gracePeriod=30 Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.758780 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c1e20b1a-3009-4bcb-b936-39aade55a6d0" containerName="nova-api-api" containerID="cri-o://f0fa717771b2744e99974d5a0a38277c3759cd0f335c2e7074450c71721a16ba" gracePeriod=30 Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.775864 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.776069 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d3705e0e-22b4-4691-b190-4173ff857f90" containerName="nova-scheduler-scheduler" containerID="cri-o://0962d98fee0e2b6d711d5b4cf85c5ab78c7ef9fa77c484483c28a583ac101230" gracePeriod=30 Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.805224 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.805774 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4f156453-e8c4-4106-bbfa-c25001c2f8e3" containerName="nova-metadata-metadata" containerID="cri-o://eca1c37f4ad609b0b529de2418979bcb6996fae3c95d43e07349a686bbf8bced" gracePeriod=30 Oct 03 16:46:51 crc kubenswrapper[4744]: I1003 16:46:51.805629 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4f156453-e8c4-4106-bbfa-c25001c2f8e3" containerName="nova-metadata-log" containerID="cri-o://943f82021ea5f083a16ea17129459a79e9148fc9ee81cc68bd5d4562e035c094" gracePeriod=30 Oct 03 16:46:52 crc kubenswrapper[4744]: I1003 16:46:52.562310 4744 generic.go:334] "Generic (PLEG): container finished" podID="c1e20b1a-3009-4bcb-b936-39aade55a6d0" containerID="0407bd8c6d71b297d98817f30c443032c1c6bf771addc7a789a8ec0cc81c0e98" exitCode=143 Oct 03 16:46:52 crc kubenswrapper[4744]: I1003 16:46:52.562833 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1e20b1a-3009-4bcb-b936-39aade55a6d0","Type":"ContainerDied","Data":"0407bd8c6d71b297d98817f30c443032c1c6bf771addc7a789a8ec0cc81c0e98"} Oct 03 16:46:52 crc kubenswrapper[4744]: I1003 16:46:52.569309 4744 generic.go:334] "Generic (PLEG): container finished" podID="4f156453-e8c4-4106-bbfa-c25001c2f8e3" containerID="943f82021ea5f083a16ea17129459a79e9148fc9ee81cc68bd5d4562e035c094" exitCode=143 Oct 03 16:46:52 crc kubenswrapper[4744]: I1003 16:46:52.569348 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4f156453-e8c4-4106-bbfa-c25001c2f8e3","Type":"ContainerDied","Data":"943f82021ea5f083a16ea17129459a79e9148fc9ee81cc68bd5d4562e035c094"} Oct 03 16:46:53 crc kubenswrapper[4744]: I1003 16:46:53.580988 4744 generic.go:334] "Generic (PLEG): container finished" podID="d3705e0e-22b4-4691-b190-4173ff857f90" containerID="0962d98fee0e2b6d711d5b4cf85c5ab78c7ef9fa77c484483c28a583ac101230" exitCode=0 Oct 03 16:46:53 crc kubenswrapper[4744]: I1003 16:46:53.581081 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d3705e0e-22b4-4691-b190-4173ff857f90","Type":"ContainerDied","Data":"0962d98fee0e2b6d711d5b4cf85c5ab78c7ef9fa77c484483c28a583ac101230"} Oct 03 16:46:53 crc kubenswrapper[4744]: I1003 16:46:53.730279 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 16:46:53 crc kubenswrapper[4744]: I1003 16:46:53.832980 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ctdr\" (UniqueName: \"kubernetes.io/projected/d3705e0e-22b4-4691-b190-4173ff857f90-kube-api-access-9ctdr\") pod \"d3705e0e-22b4-4691-b190-4173ff857f90\" (UID: \"d3705e0e-22b4-4691-b190-4173ff857f90\") " Oct 03 16:46:53 crc kubenswrapper[4744]: I1003 16:46:53.833055 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3705e0e-22b4-4691-b190-4173ff857f90-config-data\") pod \"d3705e0e-22b4-4691-b190-4173ff857f90\" (UID: \"d3705e0e-22b4-4691-b190-4173ff857f90\") " Oct 03 16:46:53 crc kubenswrapper[4744]: I1003 16:46:53.833176 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3705e0e-22b4-4691-b190-4173ff857f90-combined-ca-bundle\") pod \"d3705e0e-22b4-4691-b190-4173ff857f90\" (UID: \"d3705e0e-22b4-4691-b190-4173ff857f90\") " Oct 03 16:46:53 crc kubenswrapper[4744]: I1003 16:46:53.838729 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3705e0e-22b4-4691-b190-4173ff857f90-kube-api-access-9ctdr" (OuterVolumeSpecName: "kube-api-access-9ctdr") pod "d3705e0e-22b4-4691-b190-4173ff857f90" (UID: "d3705e0e-22b4-4691-b190-4173ff857f90"). InnerVolumeSpecName "kube-api-access-9ctdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:53 crc kubenswrapper[4744]: I1003 16:46:53.885918 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3705e0e-22b4-4691-b190-4173ff857f90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3705e0e-22b4-4691-b190-4173ff857f90" (UID: "d3705e0e-22b4-4691-b190-4173ff857f90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:53 crc kubenswrapper[4744]: I1003 16:46:53.888892 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3705e0e-22b4-4691-b190-4173ff857f90-config-data" (OuterVolumeSpecName: "config-data") pod "d3705e0e-22b4-4691-b190-4173ff857f90" (UID: "d3705e0e-22b4-4691-b190-4173ff857f90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:53 crc kubenswrapper[4744]: I1003 16:46:53.935580 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ctdr\" (UniqueName: \"kubernetes.io/projected/d3705e0e-22b4-4691-b190-4173ff857f90-kube-api-access-9ctdr\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:53 crc kubenswrapper[4744]: I1003 16:46:53.935727 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3705e0e-22b4-4691-b190-4173ff857f90-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:53 crc kubenswrapper[4744]: I1003 16:46:53.935810 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3705e0e-22b4-4691-b190-4173ff857f90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.594208 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d3705e0e-22b4-4691-b190-4173ff857f90","Type":"ContainerDied","Data":"3b7aa633beb22d32f429b98381c9b4d4f1834573a4bcab3c9ace4a65ddde055d"} Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.594406 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.594437 4744 scope.go:117] "RemoveContainer" containerID="0962d98fee0e2b6d711d5b4cf85c5ab78c7ef9fa77c484483c28a583ac101230" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.634862 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.642562 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.665024 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:46:54 crc kubenswrapper[4744]: E1003 16:46:54.665556 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2cdd571-e0fc-40c3-bf1f-40bfcf91913d" containerName="nova-manage" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.665575 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2cdd571-e0fc-40c3-bf1f-40bfcf91913d" containerName="nova-manage" Oct 03 16:46:54 crc kubenswrapper[4744]: E1003 16:46:54.665593 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3705e0e-22b4-4691-b190-4173ff857f90" containerName="nova-scheduler-scheduler" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.665601 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3705e0e-22b4-4691-b190-4173ff857f90" containerName="nova-scheduler-scheduler" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.665872 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2cdd571-e0fc-40c3-bf1f-40bfcf91913d" containerName="nova-manage" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.665895 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3705e0e-22b4-4691-b190-4173ff857f90" containerName="nova-scheduler-scheduler" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.666605 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.669747 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.695949 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.751897 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ae57486-d534-4c67-a235-218fe8293726-config-data\") pod \"nova-scheduler-0\" (UID: \"1ae57486-d534-4c67-a235-218fe8293726\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.752315 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5df22\" (UniqueName: \"kubernetes.io/projected/1ae57486-d534-4c67-a235-218fe8293726-kube-api-access-5df22\") pod \"nova-scheduler-0\" (UID: \"1ae57486-d534-4c67-a235-218fe8293726\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.752454 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ae57486-d534-4c67-a235-218fe8293726-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1ae57486-d534-4c67-a235-218fe8293726\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.854190 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5df22\" (UniqueName: \"kubernetes.io/projected/1ae57486-d534-4c67-a235-218fe8293726-kube-api-access-5df22\") pod \"nova-scheduler-0\" (UID: \"1ae57486-d534-4c67-a235-218fe8293726\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.854537 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ae57486-d534-4c67-a235-218fe8293726-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1ae57486-d534-4c67-a235-218fe8293726\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.854705 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ae57486-d534-4c67-a235-218fe8293726-config-data\") pod \"nova-scheduler-0\" (UID: \"1ae57486-d534-4c67-a235-218fe8293726\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.863524 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ae57486-d534-4c67-a235-218fe8293726-config-data\") pod \"nova-scheduler-0\" (UID: \"1ae57486-d534-4c67-a235-218fe8293726\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.870670 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ae57486-d534-4c67-a235-218fe8293726-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1ae57486-d534-4c67-a235-218fe8293726\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.883220 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5df22\" (UniqueName: \"kubernetes.io/projected/1ae57486-d534-4c67-a235-218fe8293726-kube-api-access-5df22\") pod \"nova-scheduler-0\" (UID: \"1ae57486-d534-4c67-a235-218fe8293726\") " pod="openstack/nova-scheduler-0" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.907668 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3705e0e-22b4-4691-b190-4173ff857f90" path="/var/lib/kubelet/pods/d3705e0e-22b4-4691-b190-4173ff857f90/volumes" Oct 03 16:46:54 crc kubenswrapper[4744]: I1003 16:46:54.993555 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.228257 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.364816 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cq8r\" (UniqueName: \"kubernetes.io/projected/c1e20b1a-3009-4bcb-b936-39aade55a6d0-kube-api-access-9cq8r\") pod \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.364869 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1e20b1a-3009-4bcb-b936-39aade55a6d0-logs\") pod \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.364920 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-internal-tls-certs\") pod \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.364973 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-config-data\") pod \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.365078 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-public-tls-certs\") pod \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.365142 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-combined-ca-bundle\") pod \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\" (UID: \"c1e20b1a-3009-4bcb-b936-39aade55a6d0\") " Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.365571 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1e20b1a-3009-4bcb-b936-39aade55a6d0-logs" (OuterVolumeSpecName: "logs") pod "c1e20b1a-3009-4bcb-b936-39aade55a6d0" (UID: "c1e20b1a-3009-4bcb-b936-39aade55a6d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.366812 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1e20b1a-3009-4bcb-b936-39aade55a6d0-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.378531 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1e20b1a-3009-4bcb-b936-39aade55a6d0-kube-api-access-9cq8r" (OuterVolumeSpecName: "kube-api-access-9cq8r") pod "c1e20b1a-3009-4bcb-b936-39aade55a6d0" (UID: "c1e20b1a-3009-4bcb-b936-39aade55a6d0"). InnerVolumeSpecName "kube-api-access-9cq8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.417793 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-config-data" (OuterVolumeSpecName: "config-data") pod "c1e20b1a-3009-4bcb-b936-39aade55a6d0" (UID: "c1e20b1a-3009-4bcb-b936-39aade55a6d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.420650 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1e20b1a-3009-4bcb-b936-39aade55a6d0" (UID: "c1e20b1a-3009-4bcb-b936-39aade55a6d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.453724 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c1e20b1a-3009-4bcb-b936-39aade55a6d0" (UID: "c1e20b1a-3009-4bcb-b936-39aade55a6d0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.455630 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c1e20b1a-3009-4bcb-b936-39aade55a6d0" (UID: "c1e20b1a-3009-4bcb-b936-39aade55a6d0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.469104 4744 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.469136 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.469150 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cq8r\" (UniqueName: \"kubernetes.io/projected/c1e20b1a-3009-4bcb-b936-39aade55a6d0-kube-api-access-9cq8r\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.469164 4744 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.469177 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1e20b1a-3009-4bcb-b936-39aade55a6d0-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.541157 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.543083 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 16:46:55 crc kubenswrapper[4744]: W1003 16:46:55.548473 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ae57486_d534_4c67_a235_218fe8293726.slice/crio-a5e4b2da0657c812d31670493838f8f452ffc50e93a56d17c609b01af5d705e5 WatchSource:0}: Error finding container a5e4b2da0657c812d31670493838f8f452ffc50e93a56d17c609b01af5d705e5: Status 404 returned error can't find the container with id a5e4b2da0657c812d31670493838f8f452ffc50e93a56d17c609b01af5d705e5 Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.610814 4744 generic.go:334] "Generic (PLEG): container finished" podID="4f156453-e8c4-4106-bbfa-c25001c2f8e3" containerID="eca1c37f4ad609b0b529de2418979bcb6996fae3c95d43e07349a686bbf8bced" exitCode=0 Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.610869 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4f156453-e8c4-4106-bbfa-c25001c2f8e3","Type":"ContainerDied","Data":"eca1c37f4ad609b0b529de2418979bcb6996fae3c95d43e07349a686bbf8bced"} Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.610894 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4f156453-e8c4-4106-bbfa-c25001c2f8e3","Type":"ContainerDied","Data":"bda7a1ef5c834420b06ed75c81d3167e77cd67ef9c410dc7a8b08963797445c1"} Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.610912 4744 scope.go:117] "RemoveContainer" containerID="eca1c37f4ad609b0b529de2418979bcb6996fae3c95d43e07349a686bbf8bced" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.610992 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.615403 4744 generic.go:334] "Generic (PLEG): container finished" podID="c1e20b1a-3009-4bcb-b936-39aade55a6d0" containerID="f0fa717771b2744e99974d5a0a38277c3759cd0f335c2e7074450c71721a16ba" exitCode=0 Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.615446 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.615444 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1e20b1a-3009-4bcb-b936-39aade55a6d0","Type":"ContainerDied","Data":"f0fa717771b2744e99974d5a0a38277c3759cd0f335c2e7074450c71721a16ba"} Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.615596 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1e20b1a-3009-4bcb-b936-39aade55a6d0","Type":"ContainerDied","Data":"c27c0ffbee0a6cac85b3f5454657989d60c374a431a1c308969015c18282b50a"} Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.621753 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1ae57486-d534-4c67-a235-218fe8293726","Type":"ContainerStarted","Data":"a5e4b2da0657c812d31670493838f8f452ffc50e93a56d17c609b01af5d705e5"} Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.650625 4744 scope.go:117] "RemoveContainer" containerID="943f82021ea5f083a16ea17129459a79e9148fc9ee81cc68bd5d4562e035c094" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.660487 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.675026 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-nova-metadata-tls-certs\") pod \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.675117 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-combined-ca-bundle\") pod \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.675242 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-config-data\") pod \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.675359 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8rqk\" (UniqueName: \"kubernetes.io/projected/4f156453-e8c4-4106-bbfa-c25001c2f8e3-kube-api-access-d8rqk\") pod \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.675406 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f156453-e8c4-4106-bbfa-c25001c2f8e3-logs\") pod \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\" (UID: \"4f156453-e8c4-4106-bbfa-c25001c2f8e3\") " Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.676884 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f156453-e8c4-4106-bbfa-c25001c2f8e3-logs" (OuterVolumeSpecName: "logs") pod "4f156453-e8c4-4106-bbfa-c25001c2f8e3" (UID: "4f156453-e8c4-4106-bbfa-c25001c2f8e3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.681089 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f156453-e8c4-4106-bbfa-c25001c2f8e3-kube-api-access-d8rqk" (OuterVolumeSpecName: "kube-api-access-d8rqk") pod "4f156453-e8c4-4106-bbfa-c25001c2f8e3" (UID: "4f156453-e8c4-4106-bbfa-c25001c2f8e3"). InnerVolumeSpecName "kube-api-access-d8rqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.682659 4744 scope.go:117] "RemoveContainer" containerID="eca1c37f4ad609b0b529de2418979bcb6996fae3c95d43e07349a686bbf8bced" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.684190 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.695299 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:55 crc kubenswrapper[4744]: E1003 16:46:55.695790 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1e20b1a-3009-4bcb-b936-39aade55a6d0" containerName="nova-api-log" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.695824 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1e20b1a-3009-4bcb-b936-39aade55a6d0" containerName="nova-api-log" Oct 03 16:46:55 crc kubenswrapper[4744]: E1003 16:46:55.695837 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f156453-e8c4-4106-bbfa-c25001c2f8e3" containerName="nova-metadata-metadata" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.695843 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f156453-e8c4-4106-bbfa-c25001c2f8e3" containerName="nova-metadata-metadata" Oct 03 16:46:55 crc kubenswrapper[4744]: E1003 16:46:55.695870 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1e20b1a-3009-4bcb-b936-39aade55a6d0" containerName="nova-api-api" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.695877 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1e20b1a-3009-4bcb-b936-39aade55a6d0" containerName="nova-api-api" Oct 03 16:46:55 crc kubenswrapper[4744]: E1003 16:46:55.695912 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f156453-e8c4-4106-bbfa-c25001c2f8e3" containerName="nova-metadata-log" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.695918 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f156453-e8c4-4106-bbfa-c25001c2f8e3" containerName="nova-metadata-log" Oct 03 16:46:55 crc kubenswrapper[4744]: E1003 16:46:55.696755 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eca1c37f4ad609b0b529de2418979bcb6996fae3c95d43e07349a686bbf8bced\": container with ID starting with eca1c37f4ad609b0b529de2418979bcb6996fae3c95d43e07349a686bbf8bced not found: ID does not exist" containerID="eca1c37f4ad609b0b529de2418979bcb6996fae3c95d43e07349a686bbf8bced" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.696792 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eca1c37f4ad609b0b529de2418979bcb6996fae3c95d43e07349a686bbf8bced"} err="failed to get container status \"eca1c37f4ad609b0b529de2418979bcb6996fae3c95d43e07349a686bbf8bced\": rpc error: code = NotFound desc = could not find container \"eca1c37f4ad609b0b529de2418979bcb6996fae3c95d43e07349a686bbf8bced\": container with ID starting with eca1c37f4ad609b0b529de2418979bcb6996fae3c95d43e07349a686bbf8bced not found: ID does not exist" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.696814 4744 scope.go:117] "RemoveContainer" containerID="943f82021ea5f083a16ea17129459a79e9148fc9ee81cc68bd5d4562e035c094" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.697459 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f156453-e8c4-4106-bbfa-c25001c2f8e3" containerName="nova-metadata-metadata" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.697488 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1e20b1a-3009-4bcb-b936-39aade55a6d0" containerName="nova-api-api" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.697521 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f156453-e8c4-4106-bbfa-c25001c2f8e3" containerName="nova-metadata-log" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.697531 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1e20b1a-3009-4bcb-b936-39aade55a6d0" containerName="nova-api-log" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.698700 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: E1003 16:46:55.699487 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"943f82021ea5f083a16ea17129459a79e9148fc9ee81cc68bd5d4562e035c094\": container with ID starting with 943f82021ea5f083a16ea17129459a79e9148fc9ee81cc68bd5d4562e035c094 not found: ID does not exist" containerID="943f82021ea5f083a16ea17129459a79e9148fc9ee81cc68bd5d4562e035c094" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.699533 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"943f82021ea5f083a16ea17129459a79e9148fc9ee81cc68bd5d4562e035c094"} err="failed to get container status \"943f82021ea5f083a16ea17129459a79e9148fc9ee81cc68bd5d4562e035c094\": rpc error: code = NotFound desc = could not find container \"943f82021ea5f083a16ea17129459a79e9148fc9ee81cc68bd5d4562e035c094\": container with ID starting with 943f82021ea5f083a16ea17129459a79e9148fc9ee81cc68bd5d4562e035c094 not found: ID does not exist" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.699550 4744 scope.go:117] "RemoveContainer" containerID="f0fa717771b2744e99974d5a0a38277c3759cd0f335c2e7074450c71721a16ba" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.701808 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.701999 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.706463 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.709923 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.733330 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-config-data" (OuterVolumeSpecName: "config-data") pod "4f156453-e8c4-4106-bbfa-c25001c2f8e3" (UID: "4f156453-e8c4-4106-bbfa-c25001c2f8e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.739231 4744 scope.go:117] "RemoveContainer" containerID="0407bd8c6d71b297d98817f30c443032c1c6bf771addc7a789a8ec0cc81c0e98" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.752248 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f156453-e8c4-4106-bbfa-c25001c2f8e3" (UID: "4f156453-e8c4-4106-bbfa-c25001c2f8e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.765324 4744 scope.go:117] "RemoveContainer" containerID="f0fa717771b2744e99974d5a0a38277c3759cd0f335c2e7074450c71721a16ba" Oct 03 16:46:55 crc kubenswrapper[4744]: E1003 16:46:55.765899 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0fa717771b2744e99974d5a0a38277c3759cd0f335c2e7074450c71721a16ba\": container with ID starting with f0fa717771b2744e99974d5a0a38277c3759cd0f335c2e7074450c71721a16ba not found: ID does not exist" containerID="f0fa717771b2744e99974d5a0a38277c3759cd0f335c2e7074450c71721a16ba" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.765942 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0fa717771b2744e99974d5a0a38277c3759cd0f335c2e7074450c71721a16ba"} err="failed to get container status \"f0fa717771b2744e99974d5a0a38277c3759cd0f335c2e7074450c71721a16ba\": rpc error: code = NotFound desc = could not find container \"f0fa717771b2744e99974d5a0a38277c3759cd0f335c2e7074450c71721a16ba\": container with ID starting with f0fa717771b2744e99974d5a0a38277c3759cd0f335c2e7074450c71721a16ba not found: ID does not exist" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.765966 4744 scope.go:117] "RemoveContainer" containerID="0407bd8c6d71b297d98817f30c443032c1c6bf771addc7a789a8ec0cc81c0e98" Oct 03 16:46:55 crc kubenswrapper[4744]: E1003 16:46:55.766466 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0407bd8c6d71b297d98817f30c443032c1c6bf771addc7a789a8ec0cc81c0e98\": container with ID starting with 0407bd8c6d71b297d98817f30c443032c1c6bf771addc7a789a8ec0cc81c0e98 not found: ID does not exist" containerID="0407bd8c6d71b297d98817f30c443032c1c6bf771addc7a789a8ec0cc81c0e98" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.766511 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0407bd8c6d71b297d98817f30c443032c1c6bf771addc7a789a8ec0cc81c0e98"} err="failed to get container status \"0407bd8c6d71b297d98817f30c443032c1c6bf771addc7a789a8ec0cc81c0e98\": rpc error: code = NotFound desc = could not find container \"0407bd8c6d71b297d98817f30c443032c1c6bf771addc7a789a8ec0cc81c0e98\": container with ID starting with 0407bd8c6d71b297d98817f30c443032c1c6bf771addc7a789a8ec0cc81c0e98 not found: ID does not exist" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.776009 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "4f156453-e8c4-4106-bbfa-c25001c2f8e3" (UID: "4f156453-e8c4-4106-bbfa-c25001c2f8e3"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.777525 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-config-data\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.777564 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.777611 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.777636 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb9z8\" (UniqueName: \"kubernetes.io/projected/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-kube-api-access-fb9z8\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.777657 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-logs\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.777750 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-public-tls-certs\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.777805 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.777816 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.777825 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8rqk\" (UniqueName: \"kubernetes.io/projected/4f156453-e8c4-4106-bbfa-c25001c2f8e3-kube-api-access-d8rqk\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.777835 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f156453-e8c4-4106-bbfa-c25001c2f8e3-logs\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.777846 4744 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f156453-e8c4-4106-bbfa-c25001c2f8e3-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.879800 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-public-tls-certs\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.880187 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-config-data\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.880265 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.880373 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.880475 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb9z8\" (UniqueName: \"kubernetes.io/projected/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-kube-api-access-fb9z8\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.880879 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-logs\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.881264 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-logs\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.884305 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-public-tls-certs\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.884577 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-config-data\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.884973 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.885109 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.902146 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb9z8\" (UniqueName: \"kubernetes.io/projected/5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92-kube-api-access-fb9z8\") pod \"nova-api-0\" (UID: \"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92\") " pod="openstack/nova-api-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.947167 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.964888 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.976594 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.978198 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.982079 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.982302 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 03 16:46:55 crc kubenswrapper[4744]: I1003 16:46:55.987089 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.034060 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.086410 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76b30dc4-9f12-483e-874e-6654e45e521a-config-data\") pod \"nova-metadata-0\" (UID: \"76b30dc4-9f12-483e-874e-6654e45e521a\") " pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.086575 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76b30dc4-9f12-483e-874e-6654e45e521a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"76b30dc4-9f12-483e-874e-6654e45e521a\") " pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.086610 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/76b30dc4-9f12-483e-874e-6654e45e521a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"76b30dc4-9f12-483e-874e-6654e45e521a\") " pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.086629 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq5dm\" (UniqueName: \"kubernetes.io/projected/76b30dc4-9f12-483e-874e-6654e45e521a-kube-api-access-lq5dm\") pod \"nova-metadata-0\" (UID: \"76b30dc4-9f12-483e-874e-6654e45e521a\") " pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.086657 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76b30dc4-9f12-483e-874e-6654e45e521a-logs\") pod \"nova-metadata-0\" (UID: \"76b30dc4-9f12-483e-874e-6654e45e521a\") " pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.188000 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76b30dc4-9f12-483e-874e-6654e45e521a-logs\") pod \"nova-metadata-0\" (UID: \"76b30dc4-9f12-483e-874e-6654e45e521a\") " pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.188113 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76b30dc4-9f12-483e-874e-6654e45e521a-config-data\") pod \"nova-metadata-0\" (UID: \"76b30dc4-9f12-483e-874e-6654e45e521a\") " pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.188215 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76b30dc4-9f12-483e-874e-6654e45e521a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"76b30dc4-9f12-483e-874e-6654e45e521a\") " pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.188247 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/76b30dc4-9f12-483e-874e-6654e45e521a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"76b30dc4-9f12-483e-874e-6654e45e521a\") " pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.188265 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq5dm\" (UniqueName: \"kubernetes.io/projected/76b30dc4-9f12-483e-874e-6654e45e521a-kube-api-access-lq5dm\") pod \"nova-metadata-0\" (UID: \"76b30dc4-9f12-483e-874e-6654e45e521a\") " pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.188898 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76b30dc4-9f12-483e-874e-6654e45e521a-logs\") pod \"nova-metadata-0\" (UID: \"76b30dc4-9f12-483e-874e-6654e45e521a\") " pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.192953 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76b30dc4-9f12-483e-874e-6654e45e521a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"76b30dc4-9f12-483e-874e-6654e45e521a\") " pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.192976 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76b30dc4-9f12-483e-874e-6654e45e521a-config-data\") pod \"nova-metadata-0\" (UID: \"76b30dc4-9f12-483e-874e-6654e45e521a\") " pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.193446 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/76b30dc4-9f12-483e-874e-6654e45e521a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"76b30dc4-9f12-483e-874e-6654e45e521a\") " pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.204119 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq5dm\" (UniqueName: \"kubernetes.io/projected/76b30dc4-9f12-483e-874e-6654e45e521a-kube-api-access-lq5dm\") pod \"nova-metadata-0\" (UID: \"76b30dc4-9f12-483e-874e-6654e45e521a\") " pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.301920 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.529021 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.634986 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92","Type":"ContainerStarted","Data":"eaea0c209f35e7796a6796f9585ab710f7d4bf273882f1271f80a2e4b6fb4f5e"} Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.637279 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1ae57486-d534-4c67-a235-218fe8293726","Type":"ContainerStarted","Data":"1d971884c05994515572d2e21453392c75c0a9ae9e30044e2a51eb84da59ed48"} Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.656784 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.65676993 podStartE2EDuration="2.65676993s" podCreationTimestamp="2025-10-03 16:46:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:46:56.655969209 +0000 UTC m=+1342.935845105" watchObservedRunningTime="2025-10-03 16:46:56.65676993 +0000 UTC m=+1342.936645826" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.762724 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 16:46:56 crc kubenswrapper[4744]: W1003 16:46:56.779716 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76b30dc4_9f12_483e_874e_6654e45e521a.slice/crio-6a361a0bb56634b3bed80876a8d6d3bf008c7f5ce04813b80aa168239827fee0 WatchSource:0}: Error finding container 6a361a0bb56634b3bed80876a8d6d3bf008c7f5ce04813b80aa168239827fee0: Status 404 returned error can't find the container with id 6a361a0bb56634b3bed80876a8d6d3bf008c7f5ce04813b80aa168239827fee0 Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.908668 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f156453-e8c4-4106-bbfa-c25001c2f8e3" path="/var/lib/kubelet/pods/4f156453-e8c4-4106-bbfa-c25001c2f8e3/volumes" Oct 03 16:46:56 crc kubenswrapper[4744]: I1003 16:46:56.910230 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1e20b1a-3009-4bcb-b936-39aade55a6d0" path="/var/lib/kubelet/pods/c1e20b1a-3009-4bcb-b936-39aade55a6d0/volumes" Oct 03 16:46:57 crc kubenswrapper[4744]: I1003 16:46:57.654588 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"76b30dc4-9f12-483e-874e-6654e45e521a","Type":"ContainerStarted","Data":"b084e1b0cd99ef7ffe819ff76d0ef7d300979e4c027c958439847f4336d483d1"} Oct 03 16:46:57 crc kubenswrapper[4744]: I1003 16:46:57.654890 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"76b30dc4-9f12-483e-874e-6654e45e521a","Type":"ContainerStarted","Data":"7ac406135709118c0db04b40abd93ab7f3b781a6bcaad6491e6f69cb6d9bf591"} Oct 03 16:46:57 crc kubenswrapper[4744]: I1003 16:46:57.654901 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"76b30dc4-9f12-483e-874e-6654e45e521a","Type":"ContainerStarted","Data":"6a361a0bb56634b3bed80876a8d6d3bf008c7f5ce04813b80aa168239827fee0"} Oct 03 16:46:57 crc kubenswrapper[4744]: I1003 16:46:57.657274 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92","Type":"ContainerStarted","Data":"4a4e593b258557aa699218b7211e339acf141825ea6c7f0400cc85d83096bf23"} Oct 03 16:46:57 crc kubenswrapper[4744]: I1003 16:46:57.657377 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92","Type":"ContainerStarted","Data":"5d3737ac5d4521b9184c7983174a8c238265efe41f9cb9932736d5295e2ba7a4"} Oct 03 16:46:57 crc kubenswrapper[4744]: I1003 16:46:57.678639 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.678621107 podStartE2EDuration="2.678621107s" podCreationTimestamp="2025-10-03 16:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:46:57.674093942 +0000 UTC m=+1343.953969838" watchObservedRunningTime="2025-10-03 16:46:57.678621107 +0000 UTC m=+1343.958497003" Oct 03 16:46:57 crc kubenswrapper[4744]: I1003 16:46:57.706165 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.706138666 podStartE2EDuration="2.706138666s" podCreationTimestamp="2025-10-03 16:46:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:46:57.696752888 +0000 UTC m=+1343.976628794" watchObservedRunningTime="2025-10-03 16:46:57.706138666 +0000 UTC m=+1343.986014572" Oct 03 16:46:59 crc kubenswrapper[4744]: I1003 16:46:59.994761 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 03 16:47:01 crc kubenswrapper[4744]: I1003 16:47:01.302531 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 16:47:01 crc kubenswrapper[4744]: I1003 16:47:01.302956 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 16:47:04 crc kubenswrapper[4744]: I1003 16:47:04.668798 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:47:04 crc kubenswrapper[4744]: I1003 16:47:04.669238 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:47:04 crc kubenswrapper[4744]: I1003 16:47:04.994726 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 03 16:47:05 crc kubenswrapper[4744]: I1003 16:47:05.033632 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 03 16:47:05 crc kubenswrapper[4744]: I1003 16:47:05.040764 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 03 16:47:05 crc kubenswrapper[4744]: I1003 16:47:05.820823 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 03 16:47:06 crc kubenswrapper[4744]: I1003 16:47:06.034269 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 16:47:06 crc kubenswrapper[4744]: I1003 16:47:06.034324 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 16:47:06 crc kubenswrapper[4744]: I1003 16:47:06.302370 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 16:47:06 crc kubenswrapper[4744]: I1003 16:47:06.302411 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 16:47:07 crc kubenswrapper[4744]: I1003 16:47:07.043408 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.217:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 16:47:07 crc kubenswrapper[4744]: I1003 16:47:07.052714 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.217:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 16:47:07 crc kubenswrapper[4744]: I1003 16:47:07.318663 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="76b30dc4-9f12-483e-874e-6654e45e521a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.218:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 16:47:07 crc kubenswrapper[4744]: I1003 16:47:07.319059 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="76b30dc4-9f12-483e-874e-6654e45e521a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.218:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 16:47:16 crc kubenswrapper[4744]: I1003 16:47:16.042412 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 16:47:16 crc kubenswrapper[4744]: I1003 16:47:16.043113 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 16:47:16 crc kubenswrapper[4744]: I1003 16:47:16.043477 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 16:47:16 crc kubenswrapper[4744]: I1003 16:47:16.043545 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 16:47:16 crc kubenswrapper[4744]: I1003 16:47:16.053788 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 16:47:16 crc kubenswrapper[4744]: I1003 16:47:16.055271 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 16:47:16 crc kubenswrapper[4744]: I1003 16:47:16.311173 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 03 16:47:16 crc kubenswrapper[4744]: I1003 16:47:16.312119 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 03 16:47:16 crc kubenswrapper[4744]: I1003 16:47:16.317186 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 03 16:47:16 crc kubenswrapper[4744]: I1003 16:47:16.907704 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 03 16:47:25 crc kubenswrapper[4744]: I1003 16:47:25.215192 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 16:47:26 crc kubenswrapper[4744]: I1003 16:47:26.212903 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 16:47:29 crc kubenswrapper[4744]: I1003 16:47:29.409106 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="7d21c39f-eece-4cf0-847b-c8ebc87c574d" containerName="rabbitmq" containerID="cri-o://355e3b5a1d962d3b2118f7e62c33c0d1c971a6aeb3586298b8b496211b6b4eb9" gracePeriod=604796 Oct 03 16:47:30 crc kubenswrapper[4744]: I1003 16:47:30.512523 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" containerName="rabbitmq" containerID="cri-o://c8e8da6d287f224992ddd8696bd60c618abf8af3e621a9504b2b841c36ad6bca" gracePeriod=604796 Oct 03 16:47:34 crc kubenswrapper[4744]: I1003 16:47:34.668800 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:47:34 crc kubenswrapper[4744]: I1003 16:47:34.668877 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:47:35 crc kubenswrapper[4744]: I1003 16:47:35.649052 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7d21c39f-eece-4cf0-847b-c8ebc87c574d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Oct 03 16:47:35 crc kubenswrapper[4744]: I1003 16:47:35.909960 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.008102 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.118053 4744 generic.go:334] "Generic (PLEG): container finished" podID="7d21c39f-eece-4cf0-847b-c8ebc87c574d" containerID="355e3b5a1d962d3b2118f7e62c33c0d1c971a6aeb3586298b8b496211b6b4eb9" exitCode=0 Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.118101 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d21c39f-eece-4cf0-847b-c8ebc87c574d","Type":"ContainerDied","Data":"355e3b5a1d962d3b2118f7e62c33c0d1c971a6aeb3586298b8b496211b6b4eb9"} Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.118141 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7d21c39f-eece-4cf0-847b-c8ebc87c574d","Type":"ContainerDied","Data":"4dfe06a0aafe99f6a5ab07830dda957bac86755dfd75c69bca2377b1473d75a6"} Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.118163 4744 scope.go:117] "RemoveContainer" containerID="355e3b5a1d962d3b2118f7e62c33c0d1c971a6aeb3586298b8b496211b6b4eb9" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.118181 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.157167 4744 scope.go:117] "RemoveContainer" containerID="50822061c1b75d267e0d3c6f99f1d88af5ebf5ff228151acf820e4a5ab291a81" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.175265 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-plugins\") pod \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.175307 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.175342 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d21c39f-eece-4cf0-847b-c8ebc87c574d-pod-info\") pod \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.175395 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cv2nf\" (UniqueName: \"kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-kube-api-access-cv2nf\") pod \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.175444 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-config-data\") pod \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.175507 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-server-conf\") pod \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.175541 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-tls\") pod \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.175634 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-confd\") pod \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.175662 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-erlang-cookie\") pod \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.175757 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d21c39f-eece-4cf0-847b-c8ebc87c574d-erlang-cookie-secret\") pod \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.175806 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "7d21c39f-eece-4cf0-847b-c8ebc87c574d" (UID: "7d21c39f-eece-4cf0-847b-c8ebc87c574d"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.175828 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-plugins-conf\") pod \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\" (UID: \"7d21c39f-eece-4cf0-847b-c8ebc87c574d\") " Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.176617 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "7d21c39f-eece-4cf0-847b-c8ebc87c574d" (UID: "7d21c39f-eece-4cf0-847b-c8ebc87c574d"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.176765 4744 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.176780 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.177166 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "7d21c39f-eece-4cf0-847b-c8ebc87c574d" (UID: "7d21c39f-eece-4cf0-847b-c8ebc87c574d"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.181578 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "7d21c39f-eece-4cf0-847b-c8ebc87c574d" (UID: "7d21c39f-eece-4cf0-847b-c8ebc87c574d"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.185436 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-kube-api-access-cv2nf" (OuterVolumeSpecName: "kube-api-access-cv2nf") pod "7d21c39f-eece-4cf0-847b-c8ebc87c574d" (UID: "7d21c39f-eece-4cf0-847b-c8ebc87c574d"). InnerVolumeSpecName "kube-api-access-cv2nf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.186296 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "7d21c39f-eece-4cf0-847b-c8ebc87c574d" (UID: "7d21c39f-eece-4cf0-847b-c8ebc87c574d"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.187696 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d21c39f-eece-4cf0-847b-c8ebc87c574d-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "7d21c39f-eece-4cf0-847b-c8ebc87c574d" (UID: "7d21c39f-eece-4cf0-847b-c8ebc87c574d"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.187796 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/7d21c39f-eece-4cf0-847b-c8ebc87c574d-pod-info" (OuterVolumeSpecName: "pod-info") pod "7d21c39f-eece-4cf0-847b-c8ebc87c574d" (UID: "7d21c39f-eece-4cf0-847b-c8ebc87c574d"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.212470 4744 scope.go:117] "RemoveContainer" containerID="355e3b5a1d962d3b2118f7e62c33c0d1c971a6aeb3586298b8b496211b6b4eb9" Oct 03 16:47:36 crc kubenswrapper[4744]: E1003 16:47:36.213028 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"355e3b5a1d962d3b2118f7e62c33c0d1c971a6aeb3586298b8b496211b6b4eb9\": container with ID starting with 355e3b5a1d962d3b2118f7e62c33c0d1c971a6aeb3586298b8b496211b6b4eb9 not found: ID does not exist" containerID="355e3b5a1d962d3b2118f7e62c33c0d1c971a6aeb3586298b8b496211b6b4eb9" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.213078 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"355e3b5a1d962d3b2118f7e62c33c0d1c971a6aeb3586298b8b496211b6b4eb9"} err="failed to get container status \"355e3b5a1d962d3b2118f7e62c33c0d1c971a6aeb3586298b8b496211b6b4eb9\": rpc error: code = NotFound desc = could not find container \"355e3b5a1d962d3b2118f7e62c33c0d1c971a6aeb3586298b8b496211b6b4eb9\": container with ID starting with 355e3b5a1d962d3b2118f7e62c33c0d1c971a6aeb3586298b8b496211b6b4eb9 not found: ID does not exist" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.213108 4744 scope.go:117] "RemoveContainer" containerID="50822061c1b75d267e0d3c6f99f1d88af5ebf5ff228151acf820e4a5ab291a81" Oct 03 16:47:36 crc kubenswrapper[4744]: E1003 16:47:36.213885 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50822061c1b75d267e0d3c6f99f1d88af5ebf5ff228151acf820e4a5ab291a81\": container with ID starting with 50822061c1b75d267e0d3c6f99f1d88af5ebf5ff228151acf820e4a5ab291a81 not found: ID does not exist" containerID="50822061c1b75d267e0d3c6f99f1d88af5ebf5ff228151acf820e4a5ab291a81" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.213943 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50822061c1b75d267e0d3c6f99f1d88af5ebf5ff228151acf820e4a5ab291a81"} err="failed to get container status \"50822061c1b75d267e0d3c6f99f1d88af5ebf5ff228151acf820e4a5ab291a81\": rpc error: code = NotFound desc = could not find container \"50822061c1b75d267e0d3c6f99f1d88af5ebf5ff228151acf820e4a5ab291a81\": container with ID starting with 50822061c1b75d267e0d3c6f99f1d88af5ebf5ff228151acf820e4a5ab291a81 not found: ID does not exist" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.217304 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-config-data" (OuterVolumeSpecName: "config-data") pod "7d21c39f-eece-4cf0-847b-c8ebc87c574d" (UID: "7d21c39f-eece-4cf0-847b-c8ebc87c574d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.260104 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-server-conf" (OuterVolumeSpecName: "server-conf") pod "7d21c39f-eece-4cf0-847b-c8ebc87c574d" (UID: "7d21c39f-eece-4cf0-847b-c8ebc87c574d"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.280613 4744 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7d21c39f-eece-4cf0-847b-c8ebc87c574d-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.280672 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.280683 4744 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7d21c39f-eece-4cf0-847b-c8ebc87c574d-pod-info\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.280694 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cv2nf\" (UniqueName: \"kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-kube-api-access-cv2nf\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.280710 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.280719 4744 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7d21c39f-eece-4cf0-847b-c8ebc87c574d-server-conf\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.280727 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.280739 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.315804 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "7d21c39f-eece-4cf0-847b-c8ebc87c574d" (UID: "7d21c39f-eece-4cf0-847b-c8ebc87c574d"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.325316 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.387740 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7d21c39f-eece-4cf0-847b-c8ebc87c574d-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.387765 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.464271 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.481727 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.504696 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 16:47:36 crc kubenswrapper[4744]: E1003 16:47:36.505243 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d21c39f-eece-4cf0-847b-c8ebc87c574d" containerName="rabbitmq" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.505265 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d21c39f-eece-4cf0-847b-c8ebc87c574d" containerName="rabbitmq" Oct 03 16:47:36 crc kubenswrapper[4744]: E1003 16:47:36.505311 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d21c39f-eece-4cf0-847b-c8ebc87c574d" containerName="setup-container" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.505321 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d21c39f-eece-4cf0-847b-c8ebc87c574d" containerName="setup-container" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.505590 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d21c39f-eece-4cf0-847b-c8ebc87c574d" containerName="rabbitmq" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.506900 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.509023 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.510327 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.510550 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.510759 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.510896 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.511191 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.511390 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-wjds9" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.516472 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.590823 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52159513-8a05-44c2-89f2-fb6d08e50b36-config-data\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.590875 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/52159513-8a05-44c2-89f2-fb6d08e50b36-pod-info\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.590896 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/52159513-8a05-44c2-89f2-fb6d08e50b36-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.591025 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/52159513-8a05-44c2-89f2-fb6d08e50b36-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.591064 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/52159513-8a05-44c2-89f2-fb6d08e50b36-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.591421 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/52159513-8a05-44c2-89f2-fb6d08e50b36-server-conf\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.591554 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/52159513-8a05-44c2-89f2-fb6d08e50b36-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.591602 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.591640 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/52159513-8a05-44c2-89f2-fb6d08e50b36-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.591674 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdfzb\" (UniqueName: \"kubernetes.io/projected/52159513-8a05-44c2-89f2-fb6d08e50b36-kube-api-access-wdfzb\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.591707 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/52159513-8a05-44c2-89f2-fb6d08e50b36-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.694219 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/52159513-8a05-44c2-89f2-fb6d08e50b36-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.694298 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/52159513-8a05-44c2-89f2-fb6d08e50b36-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.694322 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/52159513-8a05-44c2-89f2-fb6d08e50b36-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.694417 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/52159513-8a05-44c2-89f2-fb6d08e50b36-server-conf\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.694455 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/52159513-8a05-44c2-89f2-fb6d08e50b36-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.694486 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.694530 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/52159513-8a05-44c2-89f2-fb6d08e50b36-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.694556 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdfzb\" (UniqueName: \"kubernetes.io/projected/52159513-8a05-44c2-89f2-fb6d08e50b36-kube-api-access-wdfzb\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.694579 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/52159513-8a05-44c2-89f2-fb6d08e50b36-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.694641 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52159513-8a05-44c2-89f2-fb6d08e50b36-config-data\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.694674 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/52159513-8a05-44c2-89f2-fb6d08e50b36-pod-info\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.695053 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.695755 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/52159513-8a05-44c2-89f2-fb6d08e50b36-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.696389 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/52159513-8a05-44c2-89f2-fb6d08e50b36-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.696781 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/52159513-8a05-44c2-89f2-fb6d08e50b36-server-conf\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.697118 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/52159513-8a05-44c2-89f2-fb6d08e50b36-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.697523 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52159513-8a05-44c2-89f2-fb6d08e50b36-config-data\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.701134 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/52159513-8a05-44c2-89f2-fb6d08e50b36-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.701157 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/52159513-8a05-44c2-89f2-fb6d08e50b36-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.701243 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/52159513-8a05-44c2-89f2-fb6d08e50b36-pod-info\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.701241 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/52159513-8a05-44c2-89f2-fb6d08e50b36-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.714013 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdfzb\" (UniqueName: \"kubernetes.io/projected/52159513-8a05-44c2-89f2-fb6d08e50b36-kube-api-access-wdfzb\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.738626 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"52159513-8a05-44c2-89f2-fb6d08e50b36\") " pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.831089 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 16:47:36 crc kubenswrapper[4744]: I1003 16:47:36.923803 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d21c39f-eece-4cf0-847b-c8ebc87c574d" path="/var/lib/kubelet/pods/7d21c39f-eece-4cf0-847b-c8ebc87c574d/volumes" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.131893 4744 generic.go:334] "Generic (PLEG): container finished" podID="a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" containerID="c8e8da6d287f224992ddd8696bd60c618abf8af3e621a9504b2b841c36ad6bca" exitCode=0 Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.132087 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533","Type":"ContainerDied","Data":"c8e8da6d287f224992ddd8696bd60c618abf8af3e621a9504b2b841c36ad6bca"} Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.132148 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533","Type":"ContainerDied","Data":"fc147916d7be3ea0ecffcea13c3393b9a48fef121fe2ca2d7a881352ea627628"} Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.132168 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc147916d7be3ea0ecffcea13c3393b9a48fef121fe2ca2d7a881352ea627628" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.148006 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.204634 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-plugins-conf\") pod \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.204706 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-plugins\") pod \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.204758 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-erlang-cookie-secret\") pod \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.204797 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-confd\") pod \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.204855 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-server-conf\") pod \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.204889 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-config-data\") pod \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.204977 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hphw\" (UniqueName: \"kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-kube-api-access-6hphw\") pod \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.205004 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-erlang-cookie\") pod \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.205055 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-tls\") pod \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.205105 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.205121 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" (UID: "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.205167 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-pod-info\") pod \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\" (UID: \"a05294a3-e8e1-40cb-8e34-1c4fd4f0c533\") " Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.205749 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.206228 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" (UID: "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.206294 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" (UID: "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.211799 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-pod-info" (OuterVolumeSpecName: "pod-info") pod "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" (UID: "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.215218 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-kube-api-access-6hphw" (OuterVolumeSpecName: "kube-api-access-6hphw") pod "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" (UID: "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533"). InnerVolumeSpecName "kube-api-access-6hphw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.216653 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" (UID: "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.219457 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" (UID: "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.227863 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" (UID: "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.258312 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-config-data" (OuterVolumeSpecName: "config-data") pod "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" (UID: "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.293064 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-server-conf" (OuterVolumeSpecName: "server-conf") pod "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" (UID: "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.307380 4744 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.307412 4744 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-server-conf\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.307421 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.307430 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hphw\" (UniqueName: \"kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-kube-api-access-6hphw\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.307440 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.307448 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.307477 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.307526 4744 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-pod-info\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.307539 4744 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.335857 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.344278 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" (UID: "a05294a3-e8e1-40cb-8e34-1c4fd4f0c533"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.374376 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.409184 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:37 crc kubenswrapper[4744]: I1003 16:47:37.409213 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.143733 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.143742 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"52159513-8a05-44c2-89f2-fb6d08e50b36","Type":"ContainerStarted","Data":"38f82167f5e7bf5da7e9dadc3fa7b95c1d039f5253804ac49d65381a05f56b12"} Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.198109 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.221030 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.235342 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 16:47:38 crc kubenswrapper[4744]: E1003 16:47:38.235844 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" containerName="setup-container" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.235871 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" containerName="setup-container" Oct 03 16:47:38 crc kubenswrapper[4744]: E1003 16:47:38.235897 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" containerName="rabbitmq" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.235906 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" containerName="rabbitmq" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.236159 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" containerName="rabbitmq" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.237431 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.240658 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.240686 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.241015 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.241148 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.241248 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.241358 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-7d2qj" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.244981 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.255266 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.426735 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5d34a40f-b279-4d50-88c9-bb7776fc678b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.427176 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.427214 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5d34a40f-b279-4d50-88c9-bb7776fc678b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.427521 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5d34a40f-b279-4d50-88c9-bb7776fc678b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.427546 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqj7w\" (UniqueName: \"kubernetes.io/projected/5d34a40f-b279-4d50-88c9-bb7776fc678b-kube-api-access-qqj7w\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.427627 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5d34a40f-b279-4d50-88c9-bb7776fc678b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.427713 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5d34a40f-b279-4d50-88c9-bb7776fc678b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.427753 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5d34a40f-b279-4d50-88c9-bb7776fc678b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.427777 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5d34a40f-b279-4d50-88c9-bb7776fc678b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.427797 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d34a40f-b279-4d50-88c9-bb7776fc678b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.427859 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5d34a40f-b279-4d50-88c9-bb7776fc678b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.530109 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5d34a40f-b279-4d50-88c9-bb7776fc678b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.530190 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5d34a40f-b279-4d50-88c9-bb7776fc678b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.530233 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5d34a40f-b279-4d50-88c9-bb7776fc678b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.530253 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5d34a40f-b279-4d50-88c9-bb7776fc678b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.530287 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d34a40f-b279-4d50-88c9-bb7776fc678b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.530327 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5d34a40f-b279-4d50-88c9-bb7776fc678b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.530361 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5d34a40f-b279-4d50-88c9-bb7776fc678b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.530394 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.530422 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5d34a40f-b279-4d50-88c9-bb7776fc678b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.530443 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5d34a40f-b279-4d50-88c9-bb7776fc678b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.530470 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqj7w\" (UniqueName: \"kubernetes.io/projected/5d34a40f-b279-4d50-88c9-bb7776fc678b-kube-api-access-qqj7w\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.530953 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5d34a40f-b279-4d50-88c9-bb7776fc678b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.531264 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5d34a40f-b279-4d50-88c9-bb7776fc678b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.531388 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5d34a40f-b279-4d50-88c9-bb7776fc678b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.531599 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.531714 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5d34a40f-b279-4d50-88c9-bb7776fc678b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.532040 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d34a40f-b279-4d50-88c9-bb7776fc678b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.537225 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5d34a40f-b279-4d50-88c9-bb7776fc678b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.538091 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5d34a40f-b279-4d50-88c9-bb7776fc678b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.557789 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqj7w\" (UniqueName: \"kubernetes.io/projected/5d34a40f-b279-4d50-88c9-bb7776fc678b-kube-api-access-qqj7w\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.619301 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5d34a40f-b279-4d50-88c9-bb7776fc678b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.626025 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5d34a40f-b279-4d50-88c9-bb7776fc678b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.668090 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d34a40f-b279-4d50-88c9-bb7776fc678b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.858281 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:47:38 crc kubenswrapper[4744]: I1003 16:47:38.911057 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a05294a3-e8e1-40cb-8e34-1c4fd4f0c533" path="/var/lib/kubelet/pods/a05294a3-e8e1-40cb-8e34-1c4fd4f0c533/volumes" Oct 03 16:47:39 crc kubenswrapper[4744]: I1003 16:47:39.375194 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.160023 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d34a40f-b279-4d50-88c9-bb7776fc678b","Type":"ContainerStarted","Data":"55315c1a33e1dbe553de52a3d11159d1e7a9fac81ae457c29e752616978abf41"} Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.161603 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"52159513-8a05-44c2-89f2-fb6d08e50b36","Type":"ContainerStarted","Data":"db94de463dbcfd2d0535c395e8e96943b63b22b13037e36feece101c2c5a0497"} Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.319723 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-759799d765-dg7jt"] Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.321749 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.324022 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.334860 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-759799d765-dg7jt"] Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.479278 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-config\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.479350 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-openstack-edpm-ipam\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.479459 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-dns-svc\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.479491 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-ovsdbserver-sb\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.479532 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lw7s\" (UniqueName: \"kubernetes.io/projected/ce3ca975-d72b-4267-afba-36e045e3eb02-kube-api-access-7lw7s\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.479579 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-dns-swift-storage-0\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.479636 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-ovsdbserver-nb\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.580866 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-config\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.580928 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-openstack-edpm-ipam\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.581032 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-dns-svc\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.581061 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-ovsdbserver-sb\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.581087 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lw7s\" (UniqueName: \"kubernetes.io/projected/ce3ca975-d72b-4267-afba-36e045e3eb02-kube-api-access-7lw7s\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.581132 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-dns-swift-storage-0\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.581185 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-ovsdbserver-nb\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.582135 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-openstack-edpm-ipam\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.582139 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-ovsdbserver-sb\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.582266 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-ovsdbserver-nb\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.582329 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-dns-swift-storage-0\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.582375 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-dns-svc\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.582713 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-config\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:40 crc kubenswrapper[4744]: I1003 16:47:40.722280 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lw7s\" (UniqueName: \"kubernetes.io/projected/ce3ca975-d72b-4267-afba-36e045e3eb02-kube-api-access-7lw7s\") pod \"dnsmasq-dns-759799d765-dg7jt\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:41 crc kubenswrapper[4744]: I1003 16:47:41.017558 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:41 crc kubenswrapper[4744]: I1003 16:47:41.177684 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d34a40f-b279-4d50-88c9-bb7776fc678b","Type":"ContainerStarted","Data":"29b5e2dc3b995a1435d6110611a25bc5750aac9d0eb043a6b7851c2581b48572"} Oct 03 16:47:41 crc kubenswrapper[4744]: I1003 16:47:41.514759 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-759799d765-dg7jt"] Oct 03 16:47:42 crc kubenswrapper[4744]: I1003 16:47:42.230663 4744 generic.go:334] "Generic (PLEG): container finished" podID="ce3ca975-d72b-4267-afba-36e045e3eb02" containerID="65bef178f1c9808c718878b0226b3df32393ec849f7eb22f63748aaca5bf36ab" exitCode=0 Oct 03 16:47:42 crc kubenswrapper[4744]: I1003 16:47:42.230767 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759799d765-dg7jt" event={"ID":"ce3ca975-d72b-4267-afba-36e045e3eb02","Type":"ContainerDied","Data":"65bef178f1c9808c718878b0226b3df32393ec849f7eb22f63748aaca5bf36ab"} Oct 03 16:47:42 crc kubenswrapper[4744]: I1003 16:47:42.231399 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759799d765-dg7jt" event={"ID":"ce3ca975-d72b-4267-afba-36e045e3eb02","Type":"ContainerStarted","Data":"6a07d9e0f1cf7a2b67857405a3720afd1aa3cc7659c3ad4fccf0c11cfe4b1b95"} Oct 03 16:47:43 crc kubenswrapper[4744]: I1003 16:47:43.248875 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759799d765-dg7jt" event={"ID":"ce3ca975-d72b-4267-afba-36e045e3eb02","Type":"ContainerStarted","Data":"9a8e98bf5cc47c6ee81bcce139f809c4c5b46fad1e1962a80b0789e873465ffc"} Oct 03 16:47:43 crc kubenswrapper[4744]: I1003 16:47:43.249308 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:43 crc kubenswrapper[4744]: I1003 16:47:43.284982 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-759799d765-dg7jt" podStartSLOduration=3.284963836 podStartE2EDuration="3.284963836s" podCreationTimestamp="2025-10-03 16:47:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:47:43.275376993 +0000 UTC m=+1389.555252899" watchObservedRunningTime="2025-10-03 16:47:43.284963836 +0000 UTC m=+1389.564839722" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.019454 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.098027 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6559f4fbd7-pfrnh"] Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.098470 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" podUID="c9c81893-45c2-402f-8897-ed9d019f78b1" containerName="dnsmasq-dns" containerID="cri-o://8c1cb84ff35f8a783b54f45d869f3ba51577f5ce4d61af74f55883c8da7643ad" gracePeriod=10 Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.299750 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bb847fbb7-w6gpl"] Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.301964 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.324242 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bb847fbb7-w6gpl"] Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.347289 4744 generic.go:334] "Generic (PLEG): container finished" podID="c9c81893-45c2-402f-8897-ed9d019f78b1" containerID="8c1cb84ff35f8a783b54f45d869f3ba51577f5ce4d61af74f55883c8da7643ad" exitCode=0 Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.347341 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" event={"ID":"c9c81893-45c2-402f-8897-ed9d019f78b1","Type":"ContainerDied","Data":"8c1cb84ff35f8a783b54f45d869f3ba51577f5ce4d61af74f55883c8da7643ad"} Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.439523 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-ovsdbserver-sb\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.439679 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2kqg\" (UniqueName: \"kubernetes.io/projected/0bbdd34c-0174-4307-a779-0f8520f6e19a-kube-api-access-b2kqg\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.439808 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-openstack-edpm-ipam\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.439941 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-dns-swift-storage-0\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.439972 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-dns-svc\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.440116 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-config\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.440153 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-ovsdbserver-nb\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.543377 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-config\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.543445 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-ovsdbserver-nb\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.543475 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-ovsdbserver-sb\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.543551 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2kqg\" (UniqueName: \"kubernetes.io/projected/0bbdd34c-0174-4307-a779-0f8520f6e19a-kube-api-access-b2kqg\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.543598 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-openstack-edpm-ipam\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.543665 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-dns-swift-storage-0\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.543692 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-dns-svc\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.544481 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-dns-svc\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.545242 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-config\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.545263 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-dns-swift-storage-0\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.545390 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-ovsdbserver-nb\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.545522 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-ovsdbserver-sb\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.545519 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0bbdd34c-0174-4307-a779-0f8520f6e19a-openstack-edpm-ipam\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.570668 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2kqg\" (UniqueName: \"kubernetes.io/projected/0bbdd34c-0174-4307-a779-0f8520f6e19a-kube-api-access-b2kqg\") pod \"dnsmasq-dns-5bb847fbb7-w6gpl\" (UID: \"0bbdd34c-0174-4307-a779-0f8520f6e19a\") " pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.624305 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.765421 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.951796 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-config\") pod \"c9c81893-45c2-402f-8897-ed9d019f78b1\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.951937 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-dns-swift-storage-0\") pod \"c9c81893-45c2-402f-8897-ed9d019f78b1\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.951975 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j44tw\" (UniqueName: \"kubernetes.io/projected/c9c81893-45c2-402f-8897-ed9d019f78b1-kube-api-access-j44tw\") pod \"c9c81893-45c2-402f-8897-ed9d019f78b1\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.952037 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-ovsdbserver-sb\") pod \"c9c81893-45c2-402f-8897-ed9d019f78b1\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.952091 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-ovsdbserver-nb\") pod \"c9c81893-45c2-402f-8897-ed9d019f78b1\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.952131 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-dns-svc\") pod \"c9c81893-45c2-402f-8897-ed9d019f78b1\" (UID: \"c9c81893-45c2-402f-8897-ed9d019f78b1\") " Oct 03 16:47:51 crc kubenswrapper[4744]: I1003 16:47:51.956920 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9c81893-45c2-402f-8897-ed9d019f78b1-kube-api-access-j44tw" (OuterVolumeSpecName: "kube-api-access-j44tw") pod "c9c81893-45c2-402f-8897-ed9d019f78b1" (UID: "c9c81893-45c2-402f-8897-ed9d019f78b1"). InnerVolumeSpecName "kube-api-access-j44tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.001281 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-config" (OuterVolumeSpecName: "config") pod "c9c81893-45c2-402f-8897-ed9d019f78b1" (UID: "c9c81893-45c2-402f-8897-ed9d019f78b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.002445 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c9c81893-45c2-402f-8897-ed9d019f78b1" (UID: "c9c81893-45c2-402f-8897-ed9d019f78b1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.002571 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c9c81893-45c2-402f-8897-ed9d019f78b1" (UID: "c9c81893-45c2-402f-8897-ed9d019f78b1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.004624 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c9c81893-45c2-402f-8897-ed9d019f78b1" (UID: "c9c81893-45c2-402f-8897-ed9d019f78b1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.006631 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c9c81893-45c2-402f-8897-ed9d019f78b1" (UID: "c9c81893-45c2-402f-8897-ed9d019f78b1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.054932 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.054968 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.054982 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.054993 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.055006 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j44tw\" (UniqueName: \"kubernetes.io/projected/c9c81893-45c2-402f-8897-ed9d019f78b1-kube-api-access-j44tw\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.055019 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9c81893-45c2-402f-8897-ed9d019f78b1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.099213 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bb847fbb7-w6gpl"] Oct 03 16:47:52 crc kubenswrapper[4744]: W1003 16:47:52.099979 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0bbdd34c_0174_4307_a779_0f8520f6e19a.slice/crio-b1deea56386a834289683e45db2cc21908716b78335c666057f5525a58e34222 WatchSource:0}: Error finding container b1deea56386a834289683e45db2cc21908716b78335c666057f5525a58e34222: Status 404 returned error can't find the container with id b1deea56386a834289683e45db2cc21908716b78335c666057f5525a58e34222 Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.356741 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" event={"ID":"0bbdd34c-0174-4307-a779-0f8520f6e19a","Type":"ContainerStarted","Data":"b1deea56386a834289683e45db2cc21908716b78335c666057f5525a58e34222"} Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.358914 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" event={"ID":"c9c81893-45c2-402f-8897-ed9d019f78b1","Type":"ContainerDied","Data":"eb3636df3315e5e04417bfdc331d0837c8b44d024c9e315916ebde71e1a65eec"} Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.358956 4744 scope.go:117] "RemoveContainer" containerID="8c1cb84ff35f8a783b54f45d869f3ba51577f5ce4d61af74f55883c8da7643ad" Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.359002 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6559f4fbd7-pfrnh" Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.382820 4744 scope.go:117] "RemoveContainer" containerID="0f7ce8b158c59555f61798f11003a41c5d818ee2610f7c2906773f003db2bae5" Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.411243 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6559f4fbd7-pfrnh"] Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.420550 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6559f4fbd7-pfrnh"] Oct 03 16:47:52 crc kubenswrapper[4744]: I1003 16:47:52.903103 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9c81893-45c2-402f-8897-ed9d019f78b1" path="/var/lib/kubelet/pods/c9c81893-45c2-402f-8897-ed9d019f78b1/volumes" Oct 03 16:47:53 crc kubenswrapper[4744]: I1003 16:47:53.370287 4744 generic.go:334] "Generic (PLEG): container finished" podID="0bbdd34c-0174-4307-a779-0f8520f6e19a" containerID="974a41e237faece4976c7d379f8f51b12a30e7257d5a9aba8ac81b17645a9cbe" exitCode=0 Oct 03 16:47:53 crc kubenswrapper[4744]: I1003 16:47:53.370333 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" event={"ID":"0bbdd34c-0174-4307-a779-0f8520f6e19a","Type":"ContainerDied","Data":"974a41e237faece4976c7d379f8f51b12a30e7257d5a9aba8ac81b17645a9cbe"} Oct 03 16:47:54 crc kubenswrapper[4744]: I1003 16:47:54.384236 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" event={"ID":"0bbdd34c-0174-4307-a779-0f8520f6e19a","Type":"ContainerStarted","Data":"fd74dcff5ee9ddf2fbc52ccf3dfd554136f7647b5633dc395032c326e5fcc08b"} Oct 03 16:47:54 crc kubenswrapper[4744]: I1003 16:47:54.384686 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:47:54 crc kubenswrapper[4744]: I1003 16:47:54.417035 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" podStartSLOduration=3.4170038529999998 podStartE2EDuration="3.417003853s" podCreationTimestamp="2025-10-03 16:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:47:54.404982728 +0000 UTC m=+1400.684858704" watchObservedRunningTime="2025-10-03 16:47:54.417003853 +0000 UTC m=+1400.696879789" Oct 03 16:48:01 crc kubenswrapper[4744]: I1003 16:48:01.626520 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bb847fbb7-w6gpl" Oct 03 16:48:01 crc kubenswrapper[4744]: I1003 16:48:01.700019 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-759799d765-dg7jt"] Oct 03 16:48:01 crc kubenswrapper[4744]: I1003 16:48:01.700264 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-759799d765-dg7jt" podUID="ce3ca975-d72b-4267-afba-36e045e3eb02" containerName="dnsmasq-dns" containerID="cri-o://9a8e98bf5cc47c6ee81bcce139f809c4c5b46fad1e1962a80b0789e873465ffc" gracePeriod=10 Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.248256 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.298532 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-ovsdbserver-nb\") pod \"ce3ca975-d72b-4267-afba-36e045e3eb02\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.298740 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-config\") pod \"ce3ca975-d72b-4267-afba-36e045e3eb02\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.298769 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-ovsdbserver-sb\") pod \"ce3ca975-d72b-4267-afba-36e045e3eb02\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.298957 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-openstack-edpm-ipam\") pod \"ce3ca975-d72b-4267-afba-36e045e3eb02\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.298994 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-dns-swift-storage-0\") pod \"ce3ca975-d72b-4267-afba-36e045e3eb02\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.299030 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-dns-svc\") pod \"ce3ca975-d72b-4267-afba-36e045e3eb02\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.299087 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lw7s\" (UniqueName: \"kubernetes.io/projected/ce3ca975-d72b-4267-afba-36e045e3eb02-kube-api-access-7lw7s\") pod \"ce3ca975-d72b-4267-afba-36e045e3eb02\" (UID: \"ce3ca975-d72b-4267-afba-36e045e3eb02\") " Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.326966 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce3ca975-d72b-4267-afba-36e045e3eb02-kube-api-access-7lw7s" (OuterVolumeSpecName: "kube-api-access-7lw7s") pod "ce3ca975-d72b-4267-afba-36e045e3eb02" (UID: "ce3ca975-d72b-4267-afba-36e045e3eb02"). InnerVolumeSpecName "kube-api-access-7lw7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.363756 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ce3ca975-d72b-4267-afba-36e045e3eb02" (UID: "ce3ca975-d72b-4267-afba-36e045e3eb02"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.366780 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ce3ca975-d72b-4267-afba-36e045e3eb02" (UID: "ce3ca975-d72b-4267-afba-36e045e3eb02"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.380283 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-config" (OuterVolumeSpecName: "config") pod "ce3ca975-d72b-4267-afba-36e045e3eb02" (UID: "ce3ca975-d72b-4267-afba-36e045e3eb02"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.380333 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ce3ca975-d72b-4267-afba-36e045e3eb02" (UID: "ce3ca975-d72b-4267-afba-36e045e3eb02"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.380343 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "ce3ca975-d72b-4267-afba-36e045e3eb02" (UID: "ce3ca975-d72b-4267-afba-36e045e3eb02"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.382675 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ce3ca975-d72b-4267-afba-36e045e3eb02" (UID: "ce3ca975-d72b-4267-afba-36e045e3eb02"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.402231 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.402268 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.402283 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.402295 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lw7s\" (UniqueName: \"kubernetes.io/projected/ce3ca975-d72b-4267-afba-36e045e3eb02-kube-api-access-7lw7s\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.402308 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.402319 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-config\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.402328 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce3ca975-d72b-4267-afba-36e045e3eb02-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.485984 4744 generic.go:334] "Generic (PLEG): container finished" podID="ce3ca975-d72b-4267-afba-36e045e3eb02" containerID="9a8e98bf5cc47c6ee81bcce139f809c4c5b46fad1e1962a80b0789e873465ffc" exitCode=0 Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.486027 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759799d765-dg7jt" event={"ID":"ce3ca975-d72b-4267-afba-36e045e3eb02","Type":"ContainerDied","Data":"9a8e98bf5cc47c6ee81bcce139f809c4c5b46fad1e1962a80b0789e873465ffc"} Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.486052 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759799d765-dg7jt" event={"ID":"ce3ca975-d72b-4267-afba-36e045e3eb02","Type":"ContainerDied","Data":"6a07d9e0f1cf7a2b67857405a3720afd1aa3cc7659c3ad4fccf0c11cfe4b1b95"} Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.486069 4744 scope.go:117] "RemoveContainer" containerID="9a8e98bf5cc47c6ee81bcce139f809c4c5b46fad1e1962a80b0789e873465ffc" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.486066 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759799d765-dg7jt" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.509467 4744 scope.go:117] "RemoveContainer" containerID="65bef178f1c9808c718878b0226b3df32393ec849f7eb22f63748aaca5bf36ab" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.528809 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-759799d765-dg7jt"] Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.537001 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-759799d765-dg7jt"] Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.551056 4744 scope.go:117] "RemoveContainer" containerID="9a8e98bf5cc47c6ee81bcce139f809c4c5b46fad1e1962a80b0789e873465ffc" Oct 03 16:48:02 crc kubenswrapper[4744]: E1003 16:48:02.551400 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a8e98bf5cc47c6ee81bcce139f809c4c5b46fad1e1962a80b0789e873465ffc\": container with ID starting with 9a8e98bf5cc47c6ee81bcce139f809c4c5b46fad1e1962a80b0789e873465ffc not found: ID does not exist" containerID="9a8e98bf5cc47c6ee81bcce139f809c4c5b46fad1e1962a80b0789e873465ffc" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.551438 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a8e98bf5cc47c6ee81bcce139f809c4c5b46fad1e1962a80b0789e873465ffc"} err="failed to get container status \"9a8e98bf5cc47c6ee81bcce139f809c4c5b46fad1e1962a80b0789e873465ffc\": rpc error: code = NotFound desc = could not find container \"9a8e98bf5cc47c6ee81bcce139f809c4c5b46fad1e1962a80b0789e873465ffc\": container with ID starting with 9a8e98bf5cc47c6ee81bcce139f809c4c5b46fad1e1962a80b0789e873465ffc not found: ID does not exist" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.551465 4744 scope.go:117] "RemoveContainer" containerID="65bef178f1c9808c718878b0226b3df32393ec849f7eb22f63748aaca5bf36ab" Oct 03 16:48:02 crc kubenswrapper[4744]: E1003 16:48:02.551714 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65bef178f1c9808c718878b0226b3df32393ec849f7eb22f63748aaca5bf36ab\": container with ID starting with 65bef178f1c9808c718878b0226b3df32393ec849f7eb22f63748aaca5bf36ab not found: ID does not exist" containerID="65bef178f1c9808c718878b0226b3df32393ec849f7eb22f63748aaca5bf36ab" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.551739 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65bef178f1c9808c718878b0226b3df32393ec849f7eb22f63748aaca5bf36ab"} err="failed to get container status \"65bef178f1c9808c718878b0226b3df32393ec849f7eb22f63748aaca5bf36ab\": rpc error: code = NotFound desc = could not find container \"65bef178f1c9808c718878b0226b3df32393ec849f7eb22f63748aaca5bf36ab\": container with ID starting with 65bef178f1c9808c718878b0226b3df32393ec849f7eb22f63748aaca5bf36ab not found: ID does not exist" Oct 03 16:48:02 crc kubenswrapper[4744]: I1003 16:48:02.904326 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce3ca975-d72b-4267-afba-36e045e3eb02" path="/var/lib/kubelet/pods/ce3ca975-d72b-4267-afba-36e045e3eb02/volumes" Oct 03 16:48:04 crc kubenswrapper[4744]: I1003 16:48:04.668909 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:48:04 crc kubenswrapper[4744]: I1003 16:48:04.668973 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:48:04 crc kubenswrapper[4744]: I1003 16:48:04.669020 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:48:04 crc kubenswrapper[4744]: I1003 16:48:04.670158 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"60d0a3291947a22b9d553079ceec5bbbae302d529f9f4071b49d5de61562db0c"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 16:48:04 crc kubenswrapper[4744]: I1003 16:48:04.670233 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://60d0a3291947a22b9d553079ceec5bbbae302d529f9f4071b49d5de61562db0c" gracePeriod=600 Oct 03 16:48:05 crc kubenswrapper[4744]: I1003 16:48:05.524697 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="60d0a3291947a22b9d553079ceec5bbbae302d529f9f4071b49d5de61562db0c" exitCode=0 Oct 03 16:48:05 crc kubenswrapper[4744]: I1003 16:48:05.524836 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"60d0a3291947a22b9d553079ceec5bbbae302d529f9f4071b49d5de61562db0c"} Oct 03 16:48:05 crc kubenswrapper[4744]: I1003 16:48:05.525158 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0"} Oct 03 16:48:05 crc kubenswrapper[4744]: I1003 16:48:05.525179 4744 scope.go:117] "RemoveContainer" containerID="7b3e699f99d905cb6203d07f6a9bf0a0b414b4e9270f595c5bdaf98f6b96e23b" Oct 03 16:48:11 crc kubenswrapper[4744]: I1003 16:48:11.610545 4744 generic.go:334] "Generic (PLEG): container finished" podID="52159513-8a05-44c2-89f2-fb6d08e50b36" containerID="db94de463dbcfd2d0535c395e8e96943b63b22b13037e36feece101c2c5a0497" exitCode=0 Oct 03 16:48:11 crc kubenswrapper[4744]: I1003 16:48:11.611036 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"52159513-8a05-44c2-89f2-fb6d08e50b36","Type":"ContainerDied","Data":"db94de463dbcfd2d0535c395e8e96943b63b22b13037e36feece101c2c5a0497"} Oct 03 16:48:12 crc kubenswrapper[4744]: I1003 16:48:12.638784 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"52159513-8a05-44c2-89f2-fb6d08e50b36","Type":"ContainerStarted","Data":"06fba081bc8d8c221dfbe15b2e844a12248c2955c898380dbe8605b130cc8232"} Oct 03 16:48:12 crc kubenswrapper[4744]: I1003 16:48:12.639382 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 03 16:48:12 crc kubenswrapper[4744]: I1003 16:48:12.685642 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.685620258 podStartE2EDuration="36.685620258s" podCreationTimestamp="2025-10-03 16:47:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:48:12.670329719 +0000 UTC m=+1418.950205645" watchObservedRunningTime="2025-10-03 16:48:12.685620258 +0000 UTC m=+1418.965496174" Oct 03 16:48:13 crc kubenswrapper[4744]: I1003 16:48:13.652359 4744 generic.go:334] "Generic (PLEG): container finished" podID="5d34a40f-b279-4d50-88c9-bb7776fc678b" containerID="29b5e2dc3b995a1435d6110611a25bc5750aac9d0eb043a6b7851c2581b48572" exitCode=0 Oct 03 16:48:13 crc kubenswrapper[4744]: I1003 16:48:13.652911 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d34a40f-b279-4d50-88c9-bb7776fc678b","Type":"ContainerDied","Data":"29b5e2dc3b995a1435d6110611a25bc5750aac9d0eb043a6b7851c2581b48572"} Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.664180 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d34a40f-b279-4d50-88c9-bb7776fc678b","Type":"ContainerStarted","Data":"d751f7040bd28bcd7eb49e34b5534f096098bd32c878579869d40cb06ba20b61"} Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.665403 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.699533 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.699483391 podStartE2EDuration="36.699483391s" podCreationTimestamp="2025-10-03 16:47:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 16:48:14.690308238 +0000 UTC m=+1420.970184154" watchObservedRunningTime="2025-10-03 16:48:14.699483391 +0000 UTC m=+1420.979359307" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.725332 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7"] Oct 03 16:48:14 crc kubenswrapper[4744]: E1003 16:48:14.725735 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9c81893-45c2-402f-8897-ed9d019f78b1" containerName="init" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.725753 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9c81893-45c2-402f-8897-ed9d019f78b1" containerName="init" Oct 03 16:48:14 crc kubenswrapper[4744]: E1003 16:48:14.725776 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9c81893-45c2-402f-8897-ed9d019f78b1" containerName="dnsmasq-dns" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.725784 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9c81893-45c2-402f-8897-ed9d019f78b1" containerName="dnsmasq-dns" Oct 03 16:48:14 crc kubenswrapper[4744]: E1003 16:48:14.725804 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce3ca975-d72b-4267-afba-36e045e3eb02" containerName="init" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.725812 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce3ca975-d72b-4267-afba-36e045e3eb02" containerName="init" Oct 03 16:48:14 crc kubenswrapper[4744]: E1003 16:48:14.725832 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce3ca975-d72b-4267-afba-36e045e3eb02" containerName="dnsmasq-dns" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.725838 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce3ca975-d72b-4267-afba-36e045e3eb02" containerName="dnsmasq-dns" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.726001 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce3ca975-d72b-4267-afba-36e045e3eb02" containerName="dnsmasq-dns" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.726018 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9c81893-45c2-402f-8897-ed9d019f78b1" containerName="dnsmasq-dns" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.726614 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.731695 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.731886 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.732024 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.732136 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.747198 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7"] Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.881536 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.881630 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmd6f\" (UniqueName: \"kubernetes.io/projected/962e48ea-a515-44c9-b2c2-2e9705a31455-kube-api-access-fmd6f\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.881653 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.881861 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.984522 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.984620 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmd6f\" (UniqueName: \"kubernetes.io/projected/962e48ea-a515-44c9-b2c2-2e9705a31455-kube-api-access-fmd6f\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.984643 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.984706 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.990804 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.991127 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:14 crc kubenswrapper[4744]: I1003 16:48:14.991565 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:15 crc kubenswrapper[4744]: I1003 16:48:15.007380 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmd6f\" (UniqueName: \"kubernetes.io/projected/962e48ea-a515-44c9-b2c2-2e9705a31455-kube-api-access-fmd6f\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:15 crc kubenswrapper[4744]: I1003 16:48:15.059635 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:15 crc kubenswrapper[4744]: I1003 16:48:15.595407 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7"] Oct 03 16:48:15 crc kubenswrapper[4744]: W1003 16:48:15.597115 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod962e48ea_a515_44c9_b2c2_2e9705a31455.slice/crio-63edf0675fb139617a7c69eb6ff36e709634549ab8958ae1cae2efa0a064e384 WatchSource:0}: Error finding container 63edf0675fb139617a7c69eb6ff36e709634549ab8958ae1cae2efa0a064e384: Status 404 returned error can't find the container with id 63edf0675fb139617a7c69eb6ff36e709634549ab8958ae1cae2efa0a064e384 Oct 03 16:48:15 crc kubenswrapper[4744]: I1003 16:48:15.673056 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" event={"ID":"962e48ea-a515-44c9-b2c2-2e9705a31455","Type":"ContainerStarted","Data":"63edf0675fb139617a7c69eb6ff36e709634549ab8958ae1cae2efa0a064e384"} Oct 03 16:48:19 crc kubenswrapper[4744]: I1003 16:48:19.617450 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7hrfc"] Oct 03 16:48:19 crc kubenswrapper[4744]: I1003 16:48:19.619787 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:19 crc kubenswrapper[4744]: I1003 16:48:19.645293 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7hrfc"] Oct 03 16:48:19 crc kubenswrapper[4744]: I1003 16:48:19.789303 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ngkc\" (UniqueName: \"kubernetes.io/projected/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-kube-api-access-5ngkc\") pod \"redhat-marketplace-7hrfc\" (UID: \"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b\") " pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:19 crc kubenswrapper[4744]: I1003 16:48:19.789524 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-catalog-content\") pod \"redhat-marketplace-7hrfc\" (UID: \"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b\") " pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:19 crc kubenswrapper[4744]: I1003 16:48:19.789577 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-utilities\") pod \"redhat-marketplace-7hrfc\" (UID: \"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b\") " pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:19 crc kubenswrapper[4744]: I1003 16:48:19.890816 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-catalog-content\") pod \"redhat-marketplace-7hrfc\" (UID: \"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b\") " pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:19 crc kubenswrapper[4744]: I1003 16:48:19.890880 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-utilities\") pod \"redhat-marketplace-7hrfc\" (UID: \"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b\") " pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:19 crc kubenswrapper[4744]: I1003 16:48:19.890956 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ngkc\" (UniqueName: \"kubernetes.io/projected/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-kube-api-access-5ngkc\") pod \"redhat-marketplace-7hrfc\" (UID: \"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b\") " pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:19 crc kubenswrapper[4744]: I1003 16:48:19.891539 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-utilities\") pod \"redhat-marketplace-7hrfc\" (UID: \"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b\") " pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:19 crc kubenswrapper[4744]: I1003 16:48:19.891946 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-catalog-content\") pod \"redhat-marketplace-7hrfc\" (UID: \"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b\") " pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:19 crc kubenswrapper[4744]: I1003 16:48:19.922652 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ngkc\" (UniqueName: \"kubernetes.io/projected/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-kube-api-access-5ngkc\") pod \"redhat-marketplace-7hrfc\" (UID: \"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b\") " pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:19 crc kubenswrapper[4744]: I1003 16:48:19.950106 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:25 crc kubenswrapper[4744]: I1003 16:48:25.866795 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7hrfc"] Oct 03 16:48:26 crc kubenswrapper[4744]: I1003 16:48:26.802763 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" event={"ID":"962e48ea-a515-44c9-b2c2-2e9705a31455","Type":"ContainerStarted","Data":"e7f7289a318e76c4c1b88c0b95a6f9276cbe65abd3c93f0ef53bd7d62350db8f"} Oct 03 16:48:26 crc kubenswrapper[4744]: I1003 16:48:26.806159 4744 generic.go:334] "Generic (PLEG): container finished" podID="d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" containerID="80f5b90dc4790683446b33d5647268cccd50b7d3f9f0667868204a6eab3517c0" exitCode=0 Oct 03 16:48:26 crc kubenswrapper[4744]: I1003 16:48:26.806195 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7hrfc" event={"ID":"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b","Type":"ContainerDied","Data":"80f5b90dc4790683446b33d5647268cccd50b7d3f9f0667868204a6eab3517c0"} Oct 03 16:48:26 crc kubenswrapper[4744]: I1003 16:48:26.806215 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7hrfc" event={"ID":"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b","Type":"ContainerStarted","Data":"d0dd1a21740d1c14148154635914d2361cde6e3cd0f48cff9b51f7b6f505b8b5"} Oct 03 16:48:26 crc kubenswrapper[4744]: I1003 16:48:26.833681 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 03 16:48:26 crc kubenswrapper[4744]: I1003 16:48:26.838312 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" podStartSLOduration=2.232712568 podStartE2EDuration="12.838277005s" podCreationTimestamp="2025-10-03 16:48:14 +0000 UTC" firstStartedPulling="2025-10-03 16:48:15.599657854 +0000 UTC m=+1421.879533750" lastFinishedPulling="2025-10-03 16:48:26.205222281 +0000 UTC m=+1432.485098187" observedRunningTime="2025-10-03 16:48:26.82110128 +0000 UTC m=+1433.100977196" watchObservedRunningTime="2025-10-03 16:48:26.838277005 +0000 UTC m=+1433.118152941" Oct 03 16:48:28 crc kubenswrapper[4744]: I1003 16:48:28.826625 4744 generic.go:334] "Generic (PLEG): container finished" podID="d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" containerID="deaf42b53c1ffe83bf2215adf5352907713a94f2ec4cec98abe07c9b838ac365" exitCode=0 Oct 03 16:48:28 crc kubenswrapper[4744]: I1003 16:48:28.827155 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7hrfc" event={"ID":"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b","Type":"ContainerDied","Data":"deaf42b53c1ffe83bf2215adf5352907713a94f2ec4cec98abe07c9b838ac365"} Oct 03 16:48:28 crc kubenswrapper[4744]: I1003 16:48:28.861783 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 03 16:48:29 crc kubenswrapper[4744]: I1003 16:48:29.838014 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7hrfc" event={"ID":"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b","Type":"ContainerStarted","Data":"9ee26b79b1fe06176e92a8c1bfa4da2a880f243aa158a534faabf15e87982df0"} Oct 03 16:48:29 crc kubenswrapper[4744]: I1003 16:48:29.859903 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7hrfc" podStartSLOduration=8.433835329 podStartE2EDuration="10.859883662s" podCreationTimestamp="2025-10-03 16:48:19 +0000 UTC" firstStartedPulling="2025-10-03 16:48:26.808777217 +0000 UTC m=+1433.088653123" lastFinishedPulling="2025-10-03 16:48:29.23482556 +0000 UTC m=+1435.514701456" observedRunningTime="2025-10-03 16:48:29.855275385 +0000 UTC m=+1436.135151291" watchObservedRunningTime="2025-10-03 16:48:29.859883662 +0000 UTC m=+1436.139759558" Oct 03 16:48:29 crc kubenswrapper[4744]: I1003 16:48:29.950680 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:29 crc kubenswrapper[4744]: I1003 16:48:29.950737 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:31 crc kubenswrapper[4744]: I1003 16:48:31.001695 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-7hrfc" podUID="d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" containerName="registry-server" probeResult="failure" output=< Oct 03 16:48:31 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 03 16:48:31 crc kubenswrapper[4744]: > Oct 03 16:48:31 crc kubenswrapper[4744]: I1003 16:48:31.934916 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-74zsh"] Oct 03 16:48:31 crc kubenswrapper[4744]: I1003 16:48:31.938339 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:31 crc kubenswrapper[4744]: I1003 16:48:31.946365 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-74zsh"] Oct 03 16:48:31 crc kubenswrapper[4744]: I1003 16:48:31.968831 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmg9x\" (UniqueName: \"kubernetes.io/projected/14ae3384-3a60-4e13-a4a0-d634231409e0-kube-api-access-jmg9x\") pod \"redhat-operators-74zsh\" (UID: \"14ae3384-3a60-4e13-a4a0-d634231409e0\") " pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:31 crc kubenswrapper[4744]: I1003 16:48:31.968930 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14ae3384-3a60-4e13-a4a0-d634231409e0-utilities\") pod \"redhat-operators-74zsh\" (UID: \"14ae3384-3a60-4e13-a4a0-d634231409e0\") " pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:31 crc kubenswrapper[4744]: I1003 16:48:31.969053 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14ae3384-3a60-4e13-a4a0-d634231409e0-catalog-content\") pod \"redhat-operators-74zsh\" (UID: \"14ae3384-3a60-4e13-a4a0-d634231409e0\") " pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:32 crc kubenswrapper[4744]: I1003 16:48:32.070281 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14ae3384-3a60-4e13-a4a0-d634231409e0-catalog-content\") pod \"redhat-operators-74zsh\" (UID: \"14ae3384-3a60-4e13-a4a0-d634231409e0\") " pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:32 crc kubenswrapper[4744]: I1003 16:48:32.070436 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmg9x\" (UniqueName: \"kubernetes.io/projected/14ae3384-3a60-4e13-a4a0-d634231409e0-kube-api-access-jmg9x\") pod \"redhat-operators-74zsh\" (UID: \"14ae3384-3a60-4e13-a4a0-d634231409e0\") " pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:32 crc kubenswrapper[4744]: I1003 16:48:32.070487 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14ae3384-3a60-4e13-a4a0-d634231409e0-utilities\") pod \"redhat-operators-74zsh\" (UID: \"14ae3384-3a60-4e13-a4a0-d634231409e0\") " pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:32 crc kubenswrapper[4744]: I1003 16:48:32.070986 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14ae3384-3a60-4e13-a4a0-d634231409e0-catalog-content\") pod \"redhat-operators-74zsh\" (UID: \"14ae3384-3a60-4e13-a4a0-d634231409e0\") " pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:32 crc kubenswrapper[4744]: I1003 16:48:32.071056 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14ae3384-3a60-4e13-a4a0-d634231409e0-utilities\") pod \"redhat-operators-74zsh\" (UID: \"14ae3384-3a60-4e13-a4a0-d634231409e0\") " pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:32 crc kubenswrapper[4744]: I1003 16:48:32.090623 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmg9x\" (UniqueName: \"kubernetes.io/projected/14ae3384-3a60-4e13-a4a0-d634231409e0-kube-api-access-jmg9x\") pod \"redhat-operators-74zsh\" (UID: \"14ae3384-3a60-4e13-a4a0-d634231409e0\") " pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:32 crc kubenswrapper[4744]: I1003 16:48:32.265741 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:32 crc kubenswrapper[4744]: I1003 16:48:32.734154 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-74zsh"] Oct 03 16:48:32 crc kubenswrapper[4744]: I1003 16:48:32.865058 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74zsh" event={"ID":"14ae3384-3a60-4e13-a4a0-d634231409e0","Type":"ContainerStarted","Data":"2895efee033e18a72dd05d3deb48537033df6d6c4561b6668ed7ea26e6a57115"} Oct 03 16:48:33 crc kubenswrapper[4744]: I1003 16:48:33.876357 4744 generic.go:334] "Generic (PLEG): container finished" podID="14ae3384-3a60-4e13-a4a0-d634231409e0" containerID="29f47e222d450ba2cdeb9133710570f5aea0799133069456014d57d316731ee3" exitCode=0 Oct 03 16:48:33 crc kubenswrapper[4744]: I1003 16:48:33.876581 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74zsh" event={"ID":"14ae3384-3a60-4e13-a4a0-d634231409e0","Type":"ContainerDied","Data":"29f47e222d450ba2cdeb9133710570f5aea0799133069456014d57d316731ee3"} Oct 03 16:48:35 crc kubenswrapper[4744]: I1003 16:48:35.897565 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74zsh" event={"ID":"14ae3384-3a60-4e13-a4a0-d634231409e0","Type":"ContainerStarted","Data":"8c103a5a0f063d014822ef9037e910eb1aee0eef4f203ed1767e560dbdfc0d62"} Oct 03 16:48:36 crc kubenswrapper[4744]: I1003 16:48:36.461914 4744 scope.go:117] "RemoveContainer" containerID="8bc1ed4255353772b0f525b2c50735fff42a10b152105eb81c01847286ad05af" Oct 03 16:48:36 crc kubenswrapper[4744]: I1003 16:48:36.507675 4744 scope.go:117] "RemoveContainer" containerID="32d987bd39c8be18ca5c70041dca3bfbf8a1419d23aa8140ee93822fd9b69312" Oct 03 16:48:36 crc kubenswrapper[4744]: I1003 16:48:36.539834 4744 scope.go:117] "RemoveContainer" containerID="468e9fb03dafde6384d03462697979e76f31467f870859d2320a59286b385cf5" Oct 03 16:48:36 crc kubenswrapper[4744]: I1003 16:48:36.925698 4744 generic.go:334] "Generic (PLEG): container finished" podID="14ae3384-3a60-4e13-a4a0-d634231409e0" containerID="8c103a5a0f063d014822ef9037e910eb1aee0eef4f203ed1767e560dbdfc0d62" exitCode=0 Oct 03 16:48:36 crc kubenswrapper[4744]: I1003 16:48:36.928609 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74zsh" event={"ID":"14ae3384-3a60-4e13-a4a0-d634231409e0","Type":"ContainerDied","Data":"8c103a5a0f063d014822ef9037e910eb1aee0eef4f203ed1767e560dbdfc0d62"} Oct 03 16:48:37 crc kubenswrapper[4744]: I1003 16:48:37.944660 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74zsh" event={"ID":"14ae3384-3a60-4e13-a4a0-d634231409e0","Type":"ContainerStarted","Data":"1a28bcb1cf265b1948c5d43f2ea5c6a0c6cdb72623b3f3c332ee70f2e876c29c"} Oct 03 16:48:37 crc kubenswrapper[4744]: I1003 16:48:37.946883 4744 generic.go:334] "Generic (PLEG): container finished" podID="962e48ea-a515-44c9-b2c2-2e9705a31455" containerID="e7f7289a318e76c4c1b88c0b95a6f9276cbe65abd3c93f0ef53bd7d62350db8f" exitCode=0 Oct 03 16:48:37 crc kubenswrapper[4744]: I1003 16:48:37.946934 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" event={"ID":"962e48ea-a515-44c9-b2c2-2e9705a31455","Type":"ContainerDied","Data":"e7f7289a318e76c4c1b88c0b95a6f9276cbe65abd3c93f0ef53bd7d62350db8f"} Oct 03 16:48:37 crc kubenswrapper[4744]: I1003 16:48:37.994443 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-74zsh" podStartSLOduration=3.305488633 podStartE2EDuration="6.994420083s" podCreationTimestamp="2025-10-03 16:48:31 +0000 UTC" firstStartedPulling="2025-10-03 16:48:33.8789607 +0000 UTC m=+1440.158836606" lastFinishedPulling="2025-10-03 16:48:37.56789215 +0000 UTC m=+1443.847768056" observedRunningTime="2025-10-03 16:48:37.971300886 +0000 UTC m=+1444.251176792" watchObservedRunningTime="2025-10-03 16:48:37.994420083 +0000 UTC m=+1444.274295989" Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.418400 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.535436 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-repo-setup-combined-ca-bundle\") pod \"962e48ea-a515-44c9-b2c2-2e9705a31455\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.535711 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-ssh-key\") pod \"962e48ea-a515-44c9-b2c2-2e9705a31455\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.535738 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmd6f\" (UniqueName: \"kubernetes.io/projected/962e48ea-a515-44c9-b2c2-2e9705a31455-kube-api-access-fmd6f\") pod \"962e48ea-a515-44c9-b2c2-2e9705a31455\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.535806 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-inventory\") pod \"962e48ea-a515-44c9-b2c2-2e9705a31455\" (UID: \"962e48ea-a515-44c9-b2c2-2e9705a31455\") " Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.548404 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/962e48ea-a515-44c9-b2c2-2e9705a31455-kube-api-access-fmd6f" (OuterVolumeSpecName: "kube-api-access-fmd6f") pod "962e48ea-a515-44c9-b2c2-2e9705a31455" (UID: "962e48ea-a515-44c9-b2c2-2e9705a31455"). InnerVolumeSpecName "kube-api-access-fmd6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.561677 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "962e48ea-a515-44c9-b2c2-2e9705a31455" (UID: "962e48ea-a515-44c9-b2c2-2e9705a31455"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.575330 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "962e48ea-a515-44c9-b2c2-2e9705a31455" (UID: "962e48ea-a515-44c9-b2c2-2e9705a31455"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.594523 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-inventory" (OuterVolumeSpecName: "inventory") pod "962e48ea-a515-44c9-b2c2-2e9705a31455" (UID: "962e48ea-a515-44c9-b2c2-2e9705a31455"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.638938 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.638969 4744 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.638979 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/962e48ea-a515-44c9-b2c2-2e9705a31455-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.638990 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmd6f\" (UniqueName: \"kubernetes.io/projected/962e48ea-a515-44c9-b2c2-2e9705a31455-kube-api-access-fmd6f\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.966594 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" event={"ID":"962e48ea-a515-44c9-b2c2-2e9705a31455","Type":"ContainerDied","Data":"63edf0675fb139617a7c69eb6ff36e709634549ab8958ae1cae2efa0a064e384"} Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.966673 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63edf0675fb139617a7c69eb6ff36e709634549ab8958ae1cae2efa0a064e384" Oct 03 16:48:39 crc kubenswrapper[4744]: I1003 16:48:39.966782 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.038275 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.104880 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.122896 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9"] Oct 03 16:48:40 crc kubenswrapper[4744]: E1003 16:48:40.133529 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="962e48ea-a515-44c9-b2c2-2e9705a31455" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.133684 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="962e48ea-a515-44c9-b2c2-2e9705a31455" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.134077 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="962e48ea-a515-44c9-b2c2-2e9705a31455" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.134927 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.135552 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9"] Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.139328 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.139467 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.139790 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.143306 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.249908 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f6a23c2-45cf-4443-b1ef-befcea950ae8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tr7p9\" (UID: \"9f6a23c2-45cf-4443-b1ef-befcea950ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.250086 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfzrk\" (UniqueName: \"kubernetes.io/projected/9f6a23c2-45cf-4443-b1ef-befcea950ae8-kube-api-access-gfzrk\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tr7p9\" (UID: \"9f6a23c2-45cf-4443-b1ef-befcea950ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.250154 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f6a23c2-45cf-4443-b1ef-befcea950ae8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tr7p9\" (UID: \"9f6a23c2-45cf-4443-b1ef-befcea950ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.316222 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7hrfc"] Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.352159 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f6a23c2-45cf-4443-b1ef-befcea950ae8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tr7p9\" (UID: \"9f6a23c2-45cf-4443-b1ef-befcea950ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.352255 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f6a23c2-45cf-4443-b1ef-befcea950ae8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tr7p9\" (UID: \"9f6a23c2-45cf-4443-b1ef-befcea950ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.352406 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfzrk\" (UniqueName: \"kubernetes.io/projected/9f6a23c2-45cf-4443-b1ef-befcea950ae8-kube-api-access-gfzrk\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tr7p9\" (UID: \"9f6a23c2-45cf-4443-b1ef-befcea950ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.372837 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f6a23c2-45cf-4443-b1ef-befcea950ae8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tr7p9\" (UID: \"9f6a23c2-45cf-4443-b1ef-befcea950ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.372843 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f6a23c2-45cf-4443-b1ef-befcea950ae8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tr7p9\" (UID: \"9f6a23c2-45cf-4443-b1ef-befcea950ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.377777 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfzrk\" (UniqueName: \"kubernetes.io/projected/9f6a23c2-45cf-4443-b1ef-befcea950ae8-kube-api-access-gfzrk\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-tr7p9\" (UID: \"9f6a23c2-45cf-4443-b1ef-befcea950ae8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" Oct 03 16:48:40 crc kubenswrapper[4744]: I1003 16:48:40.458726 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" Oct 03 16:48:41 crc kubenswrapper[4744]: I1003 16:48:41.048373 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9"] Oct 03 16:48:41 crc kubenswrapper[4744]: W1003 16:48:41.055450 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f6a23c2_45cf_4443_b1ef_befcea950ae8.slice/crio-8f8639b69dea59852aa2ae0d9c5171eeb1ad1b16bda9f4bb0bc668d75820ef4d WatchSource:0}: Error finding container 8f8639b69dea59852aa2ae0d9c5171eeb1ad1b16bda9f4bb0bc668d75820ef4d: Status 404 returned error can't find the container with id 8f8639b69dea59852aa2ae0d9c5171eeb1ad1b16bda9f4bb0bc668d75820ef4d Oct 03 16:48:41 crc kubenswrapper[4744]: I1003 16:48:41.985254 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" event={"ID":"9f6a23c2-45cf-4443-b1ef-befcea950ae8","Type":"ContainerStarted","Data":"99578771ae1b2b8e3a33f6a064406bc6ca7c25cf539d73f5c63c019dd0ecbef3"} Oct 03 16:48:41 crc kubenswrapper[4744]: I1003 16:48:41.985449 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" event={"ID":"9f6a23c2-45cf-4443-b1ef-befcea950ae8","Type":"ContainerStarted","Data":"8f8639b69dea59852aa2ae0d9c5171eeb1ad1b16bda9f4bb0bc668d75820ef4d"} Oct 03 16:48:41 crc kubenswrapper[4744]: I1003 16:48:41.985556 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7hrfc" podUID="d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" containerName="registry-server" containerID="cri-o://9ee26b79b1fe06176e92a8c1bfa4da2a880f243aa158a534faabf15e87982df0" gracePeriod=2 Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.266012 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.266357 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.438435 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.463023 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" podStartSLOduration=1.864186474 podStartE2EDuration="2.463004848s" podCreationTimestamp="2025-10-03 16:48:40 +0000 UTC" firstStartedPulling="2025-10-03 16:48:41.058019496 +0000 UTC m=+1447.337895392" lastFinishedPulling="2025-10-03 16:48:41.65683786 +0000 UTC m=+1447.936713766" observedRunningTime="2025-10-03 16:48:42.0012295 +0000 UTC m=+1448.281105396" watchObservedRunningTime="2025-10-03 16:48:42.463004848 +0000 UTC m=+1448.742880744" Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.495473 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ngkc\" (UniqueName: \"kubernetes.io/projected/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-kube-api-access-5ngkc\") pod \"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b\" (UID: \"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b\") " Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.495630 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-utilities\") pod \"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b\" (UID: \"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b\") " Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.496623 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-utilities" (OuterVolumeSpecName: "utilities") pod "d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" (UID: "d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.497948 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-catalog-content\") pod \"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b\" (UID: \"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b\") " Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.498979 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.503318 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-kube-api-access-5ngkc" (OuterVolumeSpecName: "kube-api-access-5ngkc") pod "d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" (UID: "d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b"). InnerVolumeSpecName "kube-api-access-5ngkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.512185 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" (UID: "d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.601429 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.601457 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ngkc\" (UniqueName: \"kubernetes.io/projected/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b-kube-api-access-5ngkc\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.995865 4744 generic.go:334] "Generic (PLEG): container finished" podID="d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" containerID="9ee26b79b1fe06176e92a8c1bfa4da2a880f243aa158a534faabf15e87982df0" exitCode=0 Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.995920 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7hrfc" Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.995961 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7hrfc" event={"ID":"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b","Type":"ContainerDied","Data":"9ee26b79b1fe06176e92a8c1bfa4da2a880f243aa158a534faabf15e87982df0"} Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.995994 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7hrfc" event={"ID":"d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b","Type":"ContainerDied","Data":"d0dd1a21740d1c14148154635914d2361cde6e3cd0f48cff9b51f7b6f505b8b5"} Oct 03 16:48:42 crc kubenswrapper[4744]: I1003 16:48:42.996015 4744 scope.go:117] "RemoveContainer" containerID="9ee26b79b1fe06176e92a8c1bfa4da2a880f243aa158a534faabf15e87982df0" Oct 03 16:48:43 crc kubenswrapper[4744]: I1003 16:48:43.031748 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7hrfc"] Oct 03 16:48:43 crc kubenswrapper[4744]: I1003 16:48:43.032506 4744 scope.go:117] "RemoveContainer" containerID="deaf42b53c1ffe83bf2215adf5352907713a94f2ec4cec98abe07c9b838ac365" Oct 03 16:48:43 crc kubenswrapper[4744]: I1003 16:48:43.041658 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7hrfc"] Oct 03 16:48:43 crc kubenswrapper[4744]: I1003 16:48:43.054259 4744 scope.go:117] "RemoveContainer" containerID="80f5b90dc4790683446b33d5647268cccd50b7d3f9f0667868204a6eab3517c0" Oct 03 16:48:43 crc kubenswrapper[4744]: I1003 16:48:43.117986 4744 scope.go:117] "RemoveContainer" containerID="9ee26b79b1fe06176e92a8c1bfa4da2a880f243aa158a534faabf15e87982df0" Oct 03 16:48:43 crc kubenswrapper[4744]: E1003 16:48:43.118629 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ee26b79b1fe06176e92a8c1bfa4da2a880f243aa158a534faabf15e87982df0\": container with ID starting with 9ee26b79b1fe06176e92a8c1bfa4da2a880f243aa158a534faabf15e87982df0 not found: ID does not exist" containerID="9ee26b79b1fe06176e92a8c1bfa4da2a880f243aa158a534faabf15e87982df0" Oct 03 16:48:43 crc kubenswrapper[4744]: I1003 16:48:43.118762 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ee26b79b1fe06176e92a8c1bfa4da2a880f243aa158a534faabf15e87982df0"} err="failed to get container status \"9ee26b79b1fe06176e92a8c1bfa4da2a880f243aa158a534faabf15e87982df0\": rpc error: code = NotFound desc = could not find container \"9ee26b79b1fe06176e92a8c1bfa4da2a880f243aa158a534faabf15e87982df0\": container with ID starting with 9ee26b79b1fe06176e92a8c1bfa4da2a880f243aa158a534faabf15e87982df0 not found: ID does not exist" Oct 03 16:48:43 crc kubenswrapper[4744]: I1003 16:48:43.118857 4744 scope.go:117] "RemoveContainer" containerID="deaf42b53c1ffe83bf2215adf5352907713a94f2ec4cec98abe07c9b838ac365" Oct 03 16:48:43 crc kubenswrapper[4744]: E1003 16:48:43.119284 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"deaf42b53c1ffe83bf2215adf5352907713a94f2ec4cec98abe07c9b838ac365\": container with ID starting with deaf42b53c1ffe83bf2215adf5352907713a94f2ec4cec98abe07c9b838ac365 not found: ID does not exist" containerID="deaf42b53c1ffe83bf2215adf5352907713a94f2ec4cec98abe07c9b838ac365" Oct 03 16:48:43 crc kubenswrapper[4744]: I1003 16:48:43.119387 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"deaf42b53c1ffe83bf2215adf5352907713a94f2ec4cec98abe07c9b838ac365"} err="failed to get container status \"deaf42b53c1ffe83bf2215adf5352907713a94f2ec4cec98abe07c9b838ac365\": rpc error: code = NotFound desc = could not find container \"deaf42b53c1ffe83bf2215adf5352907713a94f2ec4cec98abe07c9b838ac365\": container with ID starting with deaf42b53c1ffe83bf2215adf5352907713a94f2ec4cec98abe07c9b838ac365 not found: ID does not exist" Oct 03 16:48:43 crc kubenswrapper[4744]: I1003 16:48:43.119469 4744 scope.go:117] "RemoveContainer" containerID="80f5b90dc4790683446b33d5647268cccd50b7d3f9f0667868204a6eab3517c0" Oct 03 16:48:43 crc kubenswrapper[4744]: E1003 16:48:43.119872 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80f5b90dc4790683446b33d5647268cccd50b7d3f9f0667868204a6eab3517c0\": container with ID starting with 80f5b90dc4790683446b33d5647268cccd50b7d3f9f0667868204a6eab3517c0 not found: ID does not exist" containerID="80f5b90dc4790683446b33d5647268cccd50b7d3f9f0667868204a6eab3517c0" Oct 03 16:48:43 crc kubenswrapper[4744]: I1003 16:48:43.119897 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80f5b90dc4790683446b33d5647268cccd50b7d3f9f0667868204a6eab3517c0"} err="failed to get container status \"80f5b90dc4790683446b33d5647268cccd50b7d3f9f0667868204a6eab3517c0\": rpc error: code = NotFound desc = could not find container \"80f5b90dc4790683446b33d5647268cccd50b7d3f9f0667868204a6eab3517c0\": container with ID starting with 80f5b90dc4790683446b33d5647268cccd50b7d3f9f0667868204a6eab3517c0 not found: ID does not exist" Oct 03 16:48:43 crc kubenswrapper[4744]: I1003 16:48:43.338158 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-74zsh" podUID="14ae3384-3a60-4e13-a4a0-d634231409e0" containerName="registry-server" probeResult="failure" output=< Oct 03 16:48:43 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 03 16:48:43 crc kubenswrapper[4744]: > Oct 03 16:48:44 crc kubenswrapper[4744]: I1003 16:48:44.903444 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" path="/var/lib/kubelet/pods/d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b/volumes" Oct 03 16:48:45 crc kubenswrapper[4744]: I1003 16:48:45.029068 4744 generic.go:334] "Generic (PLEG): container finished" podID="9f6a23c2-45cf-4443-b1ef-befcea950ae8" containerID="99578771ae1b2b8e3a33f6a064406bc6ca7c25cf539d73f5c63c019dd0ecbef3" exitCode=0 Oct 03 16:48:45 crc kubenswrapper[4744]: I1003 16:48:45.029116 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" event={"ID":"9f6a23c2-45cf-4443-b1ef-befcea950ae8","Type":"ContainerDied","Data":"99578771ae1b2b8e3a33f6a064406bc6ca7c25cf539d73f5c63c019dd0ecbef3"} Oct 03 16:48:46 crc kubenswrapper[4744]: I1003 16:48:46.491311 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" Oct 03 16:48:46 crc kubenswrapper[4744]: I1003 16:48:46.595746 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfzrk\" (UniqueName: \"kubernetes.io/projected/9f6a23c2-45cf-4443-b1ef-befcea950ae8-kube-api-access-gfzrk\") pod \"9f6a23c2-45cf-4443-b1ef-befcea950ae8\" (UID: \"9f6a23c2-45cf-4443-b1ef-befcea950ae8\") " Oct 03 16:48:46 crc kubenswrapper[4744]: I1003 16:48:46.595859 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f6a23c2-45cf-4443-b1ef-befcea950ae8-ssh-key\") pod \"9f6a23c2-45cf-4443-b1ef-befcea950ae8\" (UID: \"9f6a23c2-45cf-4443-b1ef-befcea950ae8\") " Oct 03 16:48:46 crc kubenswrapper[4744]: I1003 16:48:46.595905 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f6a23c2-45cf-4443-b1ef-befcea950ae8-inventory\") pod \"9f6a23c2-45cf-4443-b1ef-befcea950ae8\" (UID: \"9f6a23c2-45cf-4443-b1ef-befcea950ae8\") " Oct 03 16:48:46 crc kubenswrapper[4744]: I1003 16:48:46.602765 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f6a23c2-45cf-4443-b1ef-befcea950ae8-kube-api-access-gfzrk" (OuterVolumeSpecName: "kube-api-access-gfzrk") pod "9f6a23c2-45cf-4443-b1ef-befcea950ae8" (UID: "9f6a23c2-45cf-4443-b1ef-befcea950ae8"). InnerVolumeSpecName "kube-api-access-gfzrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:48:46 crc kubenswrapper[4744]: I1003 16:48:46.623853 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f6a23c2-45cf-4443-b1ef-befcea950ae8-inventory" (OuterVolumeSpecName: "inventory") pod "9f6a23c2-45cf-4443-b1ef-befcea950ae8" (UID: "9f6a23c2-45cf-4443-b1ef-befcea950ae8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:48:46 crc kubenswrapper[4744]: I1003 16:48:46.639571 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f6a23c2-45cf-4443-b1ef-befcea950ae8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9f6a23c2-45cf-4443-b1ef-befcea950ae8" (UID: "9f6a23c2-45cf-4443-b1ef-befcea950ae8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:48:46 crc kubenswrapper[4744]: I1003 16:48:46.698806 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f6a23c2-45cf-4443-b1ef-befcea950ae8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:46 crc kubenswrapper[4744]: I1003 16:48:46.698847 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f6a23c2-45cf-4443-b1ef-befcea950ae8-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:46 crc kubenswrapper[4744]: I1003 16:48:46.698861 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfzrk\" (UniqueName: \"kubernetes.io/projected/9f6a23c2-45cf-4443-b1ef-befcea950ae8-kube-api-access-gfzrk\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.049671 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" event={"ID":"9f6a23c2-45cf-4443-b1ef-befcea950ae8","Type":"ContainerDied","Data":"8f8639b69dea59852aa2ae0d9c5171eeb1ad1b16bda9f4bb0bc668d75820ef4d"} Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.049997 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f8639b69dea59852aa2ae0d9c5171eeb1ad1b16bda9f4bb0bc668d75820ef4d" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.049724 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-tr7p9" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.120641 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9"] Oct 03 16:48:47 crc kubenswrapper[4744]: E1003 16:48:47.121095 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" containerName="registry-server" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.121116 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" containerName="registry-server" Oct 03 16:48:47 crc kubenswrapper[4744]: E1003 16:48:47.121141 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" containerName="extract-utilities" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.121150 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" containerName="extract-utilities" Oct 03 16:48:47 crc kubenswrapper[4744]: E1003 16:48:47.121180 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" containerName="extract-content" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.121188 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" containerName="extract-content" Oct 03 16:48:47 crc kubenswrapper[4744]: E1003 16:48:47.121239 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f6a23c2-45cf-4443-b1ef-befcea950ae8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.121250 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f6a23c2-45cf-4443-b1ef-befcea950ae8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.121526 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d51a0c7b-9ee8-40cd-9838-6e62cedb6b3b" containerName="registry-server" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.121553 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f6a23c2-45cf-4443-b1ef-befcea950ae8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.122226 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.134218 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.134245 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.134224 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.134465 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.135370 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9"] Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.208109 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.208167 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.208378 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.208449 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgbp4\" (UniqueName: \"kubernetes.io/projected/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-kube-api-access-fgbp4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.310925 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.311030 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.311124 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.311180 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgbp4\" (UniqueName: \"kubernetes.io/projected/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-kube-api-access-fgbp4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.315978 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.316470 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.317695 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.332939 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgbp4\" (UniqueName: \"kubernetes.io/projected/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-kube-api-access-fgbp4\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:48:47 crc kubenswrapper[4744]: I1003 16:48:47.442027 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:48:48 crc kubenswrapper[4744]: I1003 16:48:48.026289 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9"] Oct 03 16:48:48 crc kubenswrapper[4744]: I1003 16:48:48.081320 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" event={"ID":"b860d7da-e83d-4a61-aa1d-ba22cf817d8d","Type":"ContainerStarted","Data":"8760da5c26465c1ccd0cf07025a61dcd0a1e65263ff8fbf849d2cbbd14d4bd29"} Oct 03 16:48:49 crc kubenswrapper[4744]: I1003 16:48:49.097072 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" event={"ID":"b860d7da-e83d-4a61-aa1d-ba22cf817d8d","Type":"ContainerStarted","Data":"e4729b53f2c368ac5ee02c4e222f9061a7ac807c4cacdbc506b0a5f16321aeb0"} Oct 03 16:48:49 crc kubenswrapper[4744]: I1003 16:48:49.113705 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" podStartSLOduration=1.619304451 podStartE2EDuration="2.113681256s" podCreationTimestamp="2025-10-03 16:48:47 +0000 UTC" firstStartedPulling="2025-10-03 16:48:48.026931009 +0000 UTC m=+1454.306806895" lastFinishedPulling="2025-10-03 16:48:48.521307794 +0000 UTC m=+1454.801183700" observedRunningTime="2025-10-03 16:48:49.110745311 +0000 UTC m=+1455.390621237" watchObservedRunningTime="2025-10-03 16:48:49.113681256 +0000 UTC m=+1455.393557152" Oct 03 16:48:52 crc kubenswrapper[4744]: I1003 16:48:52.333689 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:52 crc kubenswrapper[4744]: I1003 16:48:52.399849 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:52 crc kubenswrapper[4744]: I1003 16:48:52.586246 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-74zsh"] Oct 03 16:48:54 crc kubenswrapper[4744]: I1003 16:48:54.147263 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-74zsh" podUID="14ae3384-3a60-4e13-a4a0-d634231409e0" containerName="registry-server" containerID="cri-o://1a28bcb1cf265b1948c5d43f2ea5c6a0c6cdb72623b3f3c332ee70f2e876c29c" gracePeriod=2 Oct 03 16:48:54 crc kubenswrapper[4744]: I1003 16:48:54.663433 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:54 crc kubenswrapper[4744]: I1003 16:48:54.773333 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14ae3384-3a60-4e13-a4a0-d634231409e0-utilities\") pod \"14ae3384-3a60-4e13-a4a0-d634231409e0\" (UID: \"14ae3384-3a60-4e13-a4a0-d634231409e0\") " Oct 03 16:48:54 crc kubenswrapper[4744]: I1003 16:48:54.773679 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmg9x\" (UniqueName: \"kubernetes.io/projected/14ae3384-3a60-4e13-a4a0-d634231409e0-kube-api-access-jmg9x\") pod \"14ae3384-3a60-4e13-a4a0-d634231409e0\" (UID: \"14ae3384-3a60-4e13-a4a0-d634231409e0\") " Oct 03 16:48:54 crc kubenswrapper[4744]: I1003 16:48:54.773724 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14ae3384-3a60-4e13-a4a0-d634231409e0-catalog-content\") pod \"14ae3384-3a60-4e13-a4a0-d634231409e0\" (UID: \"14ae3384-3a60-4e13-a4a0-d634231409e0\") " Oct 03 16:48:54 crc kubenswrapper[4744]: I1003 16:48:54.774412 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14ae3384-3a60-4e13-a4a0-d634231409e0-utilities" (OuterVolumeSpecName: "utilities") pod "14ae3384-3a60-4e13-a4a0-d634231409e0" (UID: "14ae3384-3a60-4e13-a4a0-d634231409e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:48:54 crc kubenswrapper[4744]: I1003 16:48:54.781248 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14ae3384-3a60-4e13-a4a0-d634231409e0-kube-api-access-jmg9x" (OuterVolumeSpecName: "kube-api-access-jmg9x") pod "14ae3384-3a60-4e13-a4a0-d634231409e0" (UID: "14ae3384-3a60-4e13-a4a0-d634231409e0"). InnerVolumeSpecName "kube-api-access-jmg9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:48:54 crc kubenswrapper[4744]: I1003 16:48:54.857774 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14ae3384-3a60-4e13-a4a0-d634231409e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14ae3384-3a60-4e13-a4a0-d634231409e0" (UID: "14ae3384-3a60-4e13-a4a0-d634231409e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:48:54 crc kubenswrapper[4744]: I1003 16:48:54.876390 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14ae3384-3a60-4e13-a4a0-d634231409e0-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:54 crc kubenswrapper[4744]: I1003 16:48:54.876418 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmg9x\" (UniqueName: \"kubernetes.io/projected/14ae3384-3a60-4e13-a4a0-d634231409e0-kube-api-access-jmg9x\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:54 crc kubenswrapper[4744]: I1003 16:48:54.876432 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14ae3384-3a60-4e13-a4a0-d634231409e0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:48:55 crc kubenswrapper[4744]: I1003 16:48:55.165237 4744 generic.go:334] "Generic (PLEG): container finished" podID="14ae3384-3a60-4e13-a4a0-d634231409e0" containerID="1a28bcb1cf265b1948c5d43f2ea5c6a0c6cdb72623b3f3c332ee70f2e876c29c" exitCode=0 Oct 03 16:48:55 crc kubenswrapper[4744]: I1003 16:48:55.165294 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74zsh" event={"ID":"14ae3384-3a60-4e13-a4a0-d634231409e0","Type":"ContainerDied","Data":"1a28bcb1cf265b1948c5d43f2ea5c6a0c6cdb72623b3f3c332ee70f2e876c29c"} Oct 03 16:48:55 crc kubenswrapper[4744]: I1003 16:48:55.165326 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74zsh" Oct 03 16:48:55 crc kubenswrapper[4744]: I1003 16:48:55.165336 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74zsh" event={"ID":"14ae3384-3a60-4e13-a4a0-d634231409e0","Type":"ContainerDied","Data":"2895efee033e18a72dd05d3deb48537033df6d6c4561b6668ed7ea26e6a57115"} Oct 03 16:48:55 crc kubenswrapper[4744]: I1003 16:48:55.165351 4744 scope.go:117] "RemoveContainer" containerID="1a28bcb1cf265b1948c5d43f2ea5c6a0c6cdb72623b3f3c332ee70f2e876c29c" Oct 03 16:48:55 crc kubenswrapper[4744]: I1003 16:48:55.202093 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-74zsh"] Oct 03 16:48:55 crc kubenswrapper[4744]: I1003 16:48:55.230794 4744 scope.go:117] "RemoveContainer" containerID="8c103a5a0f063d014822ef9037e910eb1aee0eef4f203ed1767e560dbdfc0d62" Oct 03 16:48:55 crc kubenswrapper[4744]: I1003 16:48:55.252219 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-74zsh"] Oct 03 16:48:55 crc kubenswrapper[4744]: I1003 16:48:55.260601 4744 scope.go:117] "RemoveContainer" containerID="29f47e222d450ba2cdeb9133710570f5aea0799133069456014d57d316731ee3" Oct 03 16:48:55 crc kubenswrapper[4744]: I1003 16:48:55.306958 4744 scope.go:117] "RemoveContainer" containerID="1a28bcb1cf265b1948c5d43f2ea5c6a0c6cdb72623b3f3c332ee70f2e876c29c" Oct 03 16:48:55 crc kubenswrapper[4744]: E1003 16:48:55.307510 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a28bcb1cf265b1948c5d43f2ea5c6a0c6cdb72623b3f3c332ee70f2e876c29c\": container with ID starting with 1a28bcb1cf265b1948c5d43f2ea5c6a0c6cdb72623b3f3c332ee70f2e876c29c not found: ID does not exist" containerID="1a28bcb1cf265b1948c5d43f2ea5c6a0c6cdb72623b3f3c332ee70f2e876c29c" Oct 03 16:48:55 crc kubenswrapper[4744]: I1003 16:48:55.307566 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a28bcb1cf265b1948c5d43f2ea5c6a0c6cdb72623b3f3c332ee70f2e876c29c"} err="failed to get container status \"1a28bcb1cf265b1948c5d43f2ea5c6a0c6cdb72623b3f3c332ee70f2e876c29c\": rpc error: code = NotFound desc = could not find container \"1a28bcb1cf265b1948c5d43f2ea5c6a0c6cdb72623b3f3c332ee70f2e876c29c\": container with ID starting with 1a28bcb1cf265b1948c5d43f2ea5c6a0c6cdb72623b3f3c332ee70f2e876c29c not found: ID does not exist" Oct 03 16:48:55 crc kubenswrapper[4744]: I1003 16:48:55.307599 4744 scope.go:117] "RemoveContainer" containerID="8c103a5a0f063d014822ef9037e910eb1aee0eef4f203ed1767e560dbdfc0d62" Oct 03 16:48:55 crc kubenswrapper[4744]: E1003 16:48:55.308187 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c103a5a0f063d014822ef9037e910eb1aee0eef4f203ed1767e560dbdfc0d62\": container with ID starting with 8c103a5a0f063d014822ef9037e910eb1aee0eef4f203ed1767e560dbdfc0d62 not found: ID does not exist" containerID="8c103a5a0f063d014822ef9037e910eb1aee0eef4f203ed1767e560dbdfc0d62" Oct 03 16:48:55 crc kubenswrapper[4744]: I1003 16:48:55.308226 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c103a5a0f063d014822ef9037e910eb1aee0eef4f203ed1767e560dbdfc0d62"} err="failed to get container status \"8c103a5a0f063d014822ef9037e910eb1aee0eef4f203ed1767e560dbdfc0d62\": rpc error: code = NotFound desc = could not find container \"8c103a5a0f063d014822ef9037e910eb1aee0eef4f203ed1767e560dbdfc0d62\": container with ID starting with 8c103a5a0f063d014822ef9037e910eb1aee0eef4f203ed1767e560dbdfc0d62 not found: ID does not exist" Oct 03 16:48:55 crc kubenswrapper[4744]: I1003 16:48:55.308261 4744 scope.go:117] "RemoveContainer" containerID="29f47e222d450ba2cdeb9133710570f5aea0799133069456014d57d316731ee3" Oct 03 16:48:55 crc kubenswrapper[4744]: E1003 16:48:55.308841 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29f47e222d450ba2cdeb9133710570f5aea0799133069456014d57d316731ee3\": container with ID starting with 29f47e222d450ba2cdeb9133710570f5aea0799133069456014d57d316731ee3 not found: ID does not exist" containerID="29f47e222d450ba2cdeb9133710570f5aea0799133069456014d57d316731ee3" Oct 03 16:48:55 crc kubenswrapper[4744]: I1003 16:48:55.308874 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29f47e222d450ba2cdeb9133710570f5aea0799133069456014d57d316731ee3"} err="failed to get container status \"29f47e222d450ba2cdeb9133710570f5aea0799133069456014d57d316731ee3\": rpc error: code = NotFound desc = could not find container \"29f47e222d450ba2cdeb9133710570f5aea0799133069456014d57d316731ee3\": container with ID starting with 29f47e222d450ba2cdeb9133710570f5aea0799133069456014d57d316731ee3 not found: ID does not exist" Oct 03 16:48:56 crc kubenswrapper[4744]: I1003 16:48:56.908689 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14ae3384-3a60-4e13-a4a0-d634231409e0" path="/var/lib/kubelet/pods/14ae3384-3a60-4e13-a4a0-d634231409e0/volumes" Oct 03 16:49:36 crc kubenswrapper[4744]: I1003 16:49:36.952818 4744 scope.go:117] "RemoveContainer" containerID="363010deed7d9b35da3eedc23cf2ced9cec77ca56f9bcfd98c2f8a957a5a6b26" Oct 03 16:49:37 crc kubenswrapper[4744]: I1003 16:49:37.001602 4744 scope.go:117] "RemoveContainer" containerID="c8e8da6d287f224992ddd8696bd60c618abf8af3e621a9504b2b841c36ad6bca" Oct 03 16:49:37 crc kubenswrapper[4744]: I1003 16:49:37.068088 4744 scope.go:117] "RemoveContainer" containerID="1a78d69196e0287df60bf4c840b2a9b6309794774bad9c0eee264bc8bfa2c004" Oct 03 16:50:34 crc kubenswrapper[4744]: I1003 16:50:34.668285 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:50:34 crc kubenswrapper[4744]: I1003 16:50:34.669097 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:50:37 crc kubenswrapper[4744]: I1003 16:50:37.210145 4744 scope.go:117] "RemoveContainer" containerID="64e1d8fa67bd4b27a329c86f45dd74b6928ba9f3ae1ce41f3a1deadae7090fa2" Oct 03 16:50:37 crc kubenswrapper[4744]: I1003 16:50:37.235553 4744 scope.go:117] "RemoveContainer" containerID="a715f005eeddefc330647709836f3e4428fd6bf79cdba4079f4dda6559cfcf68" Oct 03 16:51:04 crc kubenswrapper[4744]: I1003 16:51:04.668231 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:51:04 crc kubenswrapper[4744]: I1003 16:51:04.668893 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:51:34 crc kubenswrapper[4744]: I1003 16:51:34.668957 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:51:34 crc kubenswrapper[4744]: I1003 16:51:34.669768 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:51:34 crc kubenswrapper[4744]: I1003 16:51:34.669840 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 16:51:34 crc kubenswrapper[4744]: I1003 16:51:34.671144 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 16:51:34 crc kubenswrapper[4744]: I1003 16:51:34.671282 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" gracePeriod=600 Oct 03 16:51:34 crc kubenswrapper[4744]: E1003 16:51:34.806181 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:51:35 crc kubenswrapper[4744]: I1003 16:51:35.188827 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0"} Oct 03 16:51:35 crc kubenswrapper[4744]: I1003 16:51:35.188888 4744 scope.go:117] "RemoveContainer" containerID="60d0a3291947a22b9d553079ceec5bbbae302d529f9f4071b49d5de61562db0c" Oct 03 16:51:35 crc kubenswrapper[4744]: I1003 16:51:35.188870 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" exitCode=0 Oct 03 16:51:35 crc kubenswrapper[4744]: I1003 16:51:35.190408 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:51:35 crc kubenswrapper[4744]: E1003 16:51:35.191333 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:51:37 crc kubenswrapper[4744]: I1003 16:51:37.274459 4744 scope.go:117] "RemoveContainer" containerID="ba6406170f7106c9dba94384d81f63d45eb970af3d62d53c86e31b1dc85d47ff" Oct 03 16:51:37 crc kubenswrapper[4744]: I1003 16:51:37.296805 4744 scope.go:117] "RemoveContainer" containerID="dc43d1e8772042d61a68de1f2b1c1c2ab881e61eac2e180894cd70a3c7153adf" Oct 03 16:51:37 crc kubenswrapper[4744]: I1003 16:51:37.318180 4744 scope.go:117] "RemoveContainer" containerID="6292a918a1fbf56b48a41e7b1d948dba59c8a26c8e38e368dfb1779a3c68d8f2" Oct 03 16:51:37 crc kubenswrapper[4744]: I1003 16:51:37.337465 4744 scope.go:117] "RemoveContainer" containerID="53ac3655289c7648b5a9b1b7f77edd67983324dc05b6862f0ec695326ada3bb9" Oct 03 16:51:45 crc kubenswrapper[4744]: I1003 16:51:45.326418 4744 generic.go:334] "Generic (PLEG): container finished" podID="b860d7da-e83d-4a61-aa1d-ba22cf817d8d" containerID="e4729b53f2c368ac5ee02c4e222f9061a7ac807c4cacdbc506b0a5f16321aeb0" exitCode=0 Oct 03 16:51:45 crc kubenswrapper[4744]: I1003 16:51:45.326549 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" event={"ID":"b860d7da-e83d-4a61-aa1d-ba22cf817d8d","Type":"ContainerDied","Data":"e4729b53f2c368ac5ee02c4e222f9061a7ac807c4cacdbc506b0a5f16321aeb0"} Oct 03 16:51:46 crc kubenswrapper[4744]: I1003 16:51:46.844274 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:51:46 crc kubenswrapper[4744]: I1003 16:51:46.893692 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:51:46 crc kubenswrapper[4744]: E1003 16:51:46.894324 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:51:46 crc kubenswrapper[4744]: I1003 16:51:46.921545 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-inventory\") pod \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " Oct 03 16:51:46 crc kubenswrapper[4744]: I1003 16:51:46.921636 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-bootstrap-combined-ca-bundle\") pod \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " Oct 03 16:51:46 crc kubenswrapper[4744]: I1003 16:51:46.922663 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-ssh-key\") pod \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " Oct 03 16:51:46 crc kubenswrapper[4744]: I1003 16:51:46.922704 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgbp4\" (UniqueName: \"kubernetes.io/projected/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-kube-api-access-fgbp4\") pod \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\" (UID: \"b860d7da-e83d-4a61-aa1d-ba22cf817d8d\") " Oct 03 16:51:46 crc kubenswrapper[4744]: I1003 16:51:46.927613 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b860d7da-e83d-4a61-aa1d-ba22cf817d8d" (UID: "b860d7da-e83d-4a61-aa1d-ba22cf817d8d"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:51:46 crc kubenswrapper[4744]: I1003 16:51:46.930911 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-kube-api-access-fgbp4" (OuterVolumeSpecName: "kube-api-access-fgbp4") pod "b860d7da-e83d-4a61-aa1d-ba22cf817d8d" (UID: "b860d7da-e83d-4a61-aa1d-ba22cf817d8d"). InnerVolumeSpecName "kube-api-access-fgbp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:51:46 crc kubenswrapper[4744]: I1003 16:51:46.954211 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-inventory" (OuterVolumeSpecName: "inventory") pod "b860d7da-e83d-4a61-aa1d-ba22cf817d8d" (UID: "b860d7da-e83d-4a61-aa1d-ba22cf817d8d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:51:46 crc kubenswrapper[4744]: I1003 16:51:46.960949 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b860d7da-e83d-4a61-aa1d-ba22cf817d8d" (UID: "b860d7da-e83d-4a61-aa1d-ba22cf817d8d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.026163 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.028221 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgbp4\" (UniqueName: \"kubernetes.io/projected/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-kube-api-access-fgbp4\") on node \"crc\" DevicePath \"\"" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.028246 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.028265 4744 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b860d7da-e83d-4a61-aa1d-ba22cf817d8d-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.353067 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" event={"ID":"b860d7da-e83d-4a61-aa1d-ba22cf817d8d","Type":"ContainerDied","Data":"8760da5c26465c1ccd0cf07025a61dcd0a1e65263ff8fbf849d2cbbd14d4bd29"} Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.353115 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.353133 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8760da5c26465c1ccd0cf07025a61dcd0a1e65263ff8fbf849d2cbbd14d4bd29" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.450926 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8"] Oct 03 16:51:47 crc kubenswrapper[4744]: E1003 16:51:47.451427 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ae3384-3a60-4e13-a4a0-d634231409e0" containerName="extract-utilities" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.451450 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ae3384-3a60-4e13-a4a0-d634231409e0" containerName="extract-utilities" Oct 03 16:51:47 crc kubenswrapper[4744]: E1003 16:51:47.451473 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ae3384-3a60-4e13-a4a0-d634231409e0" containerName="registry-server" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.451482 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ae3384-3a60-4e13-a4a0-d634231409e0" containerName="registry-server" Oct 03 16:51:47 crc kubenswrapper[4744]: E1003 16:51:47.451530 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ae3384-3a60-4e13-a4a0-d634231409e0" containerName="extract-content" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.451540 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ae3384-3a60-4e13-a4a0-d634231409e0" containerName="extract-content" Oct 03 16:51:47 crc kubenswrapper[4744]: E1003 16:51:47.451559 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b860d7da-e83d-4a61-aa1d-ba22cf817d8d" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.451569 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b860d7da-e83d-4a61-aa1d-ba22cf817d8d" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.451811 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b860d7da-e83d-4a61-aa1d-ba22cf817d8d" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.451836 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ae3384-3a60-4e13-a4a0-d634231409e0" containerName="registry-server" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.452621 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.456197 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.456290 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.456452 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.458051 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.476115 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8"] Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.541126 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8\" (UID: \"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.541380 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48zx5\" (UniqueName: \"kubernetes.io/projected/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-kube-api-access-48zx5\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8\" (UID: \"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.541478 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8\" (UID: \"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.643735 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48zx5\" (UniqueName: \"kubernetes.io/projected/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-kube-api-access-48zx5\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8\" (UID: \"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.643870 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8\" (UID: \"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.644003 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8\" (UID: \"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.649058 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8\" (UID: \"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.653357 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8\" (UID: \"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.660974 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48zx5\" (UniqueName: \"kubernetes.io/projected/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-kube-api-access-48zx5\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8\" (UID: \"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" Oct 03 16:51:47 crc kubenswrapper[4744]: I1003 16:51:47.791094 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" Oct 03 16:51:48 crc kubenswrapper[4744]: I1003 16:51:48.377723 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8"] Oct 03 16:51:48 crc kubenswrapper[4744]: W1003 16:51:48.384694 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2145bd1_f9de_41e1_b1c5_39f0dcd0c8a1.slice/crio-01632ffddcb9cf44ce14db1738c5cbbe313048c43ecbe0366341f7cf54ebd341 WatchSource:0}: Error finding container 01632ffddcb9cf44ce14db1738c5cbbe313048c43ecbe0366341f7cf54ebd341: Status 404 returned error can't find the container with id 01632ffddcb9cf44ce14db1738c5cbbe313048c43ecbe0366341f7cf54ebd341 Oct 03 16:51:48 crc kubenswrapper[4744]: I1003 16:51:48.386696 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 16:51:49 crc kubenswrapper[4744]: I1003 16:51:49.378746 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" event={"ID":"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1","Type":"ContainerStarted","Data":"58b5309126059f9585226736dab9306fdc164b56b0e50872320b4d693fbc7009"} Oct 03 16:51:49 crc kubenswrapper[4744]: I1003 16:51:49.379298 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" event={"ID":"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1","Type":"ContainerStarted","Data":"01632ffddcb9cf44ce14db1738c5cbbe313048c43ecbe0366341f7cf54ebd341"} Oct 03 16:51:49 crc kubenswrapper[4744]: I1003 16:51:49.401003 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" podStartSLOduration=1.959828337 podStartE2EDuration="2.400976116s" podCreationTimestamp="2025-10-03 16:51:47 +0000 UTC" firstStartedPulling="2025-10-03 16:51:48.386369016 +0000 UTC m=+1634.666244912" lastFinishedPulling="2025-10-03 16:51:48.827516795 +0000 UTC m=+1635.107392691" observedRunningTime="2025-10-03 16:51:49.395972483 +0000 UTC m=+1635.675848389" watchObservedRunningTime="2025-10-03 16:51:49.400976116 +0000 UTC m=+1635.680852012" Oct 03 16:51:58 crc kubenswrapper[4744]: I1003 16:51:58.892090 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:51:58 crc kubenswrapper[4744]: E1003 16:51:58.892980 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:52:13 crc kubenswrapper[4744]: I1003 16:52:13.893166 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:52:13 crc kubenswrapper[4744]: E1003 16:52:13.894329 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:52:23 crc kubenswrapper[4744]: I1003 16:52:23.040429 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-mj4kq"] Oct 03 16:52:23 crc kubenswrapper[4744]: I1003 16:52:23.047461 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-mj4kq"] Oct 03 16:52:24 crc kubenswrapper[4744]: I1003 16:52:24.910170 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6f99e55-13bd-4d45-8ab3-2e4c06485dbe" path="/var/lib/kubelet/pods/b6f99e55-13bd-4d45-8ab3-2e4c06485dbe/volumes" Oct 03 16:52:27 crc kubenswrapper[4744]: I1003 16:52:27.892579 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:52:27 crc kubenswrapper[4744]: E1003 16:52:27.893485 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:52:30 crc kubenswrapper[4744]: I1003 16:52:30.045409 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-pmltl"] Oct 03 16:52:30 crc kubenswrapper[4744]: I1003 16:52:30.056236 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-pmltl"] Oct 03 16:52:30 crc kubenswrapper[4744]: I1003 16:52:30.909228 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef2c213d-480e-4470-b69b-890209b5a6fb" path="/var/lib/kubelet/pods/ef2c213d-480e-4470-b69b-890209b5a6fb/volumes" Oct 03 16:52:32 crc kubenswrapper[4744]: I1003 16:52:32.042423 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-x55rd"] Oct 03 16:52:32 crc kubenswrapper[4744]: I1003 16:52:32.062190 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-x55rd"] Oct 03 16:52:32 crc kubenswrapper[4744]: I1003 16:52:32.911180 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f0066c0-388f-4281-83f4-1fe5b0907d1c" path="/var/lib/kubelet/pods/2f0066c0-388f-4281-83f4-1fe5b0907d1c/volumes" Oct 03 16:52:33 crc kubenswrapper[4744]: I1003 16:52:33.039420 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-cc39-account-create-c9g7l"] Oct 03 16:52:33 crc kubenswrapper[4744]: I1003 16:52:33.053076 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-cc39-account-create-c9g7l"] Oct 03 16:52:34 crc kubenswrapper[4744]: I1003 16:52:34.911404 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a08b81c-0ca0-422e-8323-6e8ba91d1a31" path="/var/lib/kubelet/pods/0a08b81c-0ca0-422e-8323-6e8ba91d1a31/volumes" Oct 03 16:52:37 crc kubenswrapper[4744]: I1003 16:52:37.403676 4744 scope.go:117] "RemoveContainer" containerID="97fc922c13c202a20cfce82ada09283fd345b0c4b6eb06d74184978d86796d28" Oct 03 16:52:37 crc kubenswrapper[4744]: I1003 16:52:37.431850 4744 scope.go:117] "RemoveContainer" containerID="223036cb094324614faa919b4300eafff5e384637710226c1ca30e1ef4854d2e" Oct 03 16:52:37 crc kubenswrapper[4744]: I1003 16:52:37.483648 4744 scope.go:117] "RemoveContainer" containerID="e78967dddcc3d21e07e0bd426dd342ffd1f6530eb46d46212a740ca2d5c0baab" Oct 03 16:52:37 crc kubenswrapper[4744]: I1003 16:52:37.524233 4744 scope.go:117] "RemoveContainer" containerID="bd91db54296bad89568c91d52e226066828dcab4426f18c5920f65119bcf722f" Oct 03 16:52:37 crc kubenswrapper[4744]: I1003 16:52:37.560611 4744 scope.go:117] "RemoveContainer" containerID="2d3ec7a1b23df0aeefde51845f97da1a41107b85563c69c240d72fdd219d0625" Oct 03 16:52:37 crc kubenswrapper[4744]: I1003 16:52:37.588205 4744 scope.go:117] "RemoveContainer" containerID="d6ac5af47e78703484a6564224485d108b26016a512fe43491200033e1079231" Oct 03 16:52:37 crc kubenswrapper[4744]: I1003 16:52:37.614591 4744 scope.go:117] "RemoveContainer" containerID="7465f321030bfc2785ae6ebb64ef1c046057a3a288da59ea404f41de37e25aae" Oct 03 16:52:38 crc kubenswrapper[4744]: I1003 16:52:38.892110 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:52:38 crc kubenswrapper[4744]: E1003 16:52:38.892788 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:52:42 crc kubenswrapper[4744]: I1003 16:52:42.053583 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-d48a-account-create-t4tk2"] Oct 03 16:52:42 crc kubenswrapper[4744]: I1003 16:52:42.064326 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-d48a-account-create-t4tk2"] Oct 03 16:52:42 crc kubenswrapper[4744]: I1003 16:52:42.906070 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5" path="/var/lib/kubelet/pods/c57a4f80-7a8e-4f2b-9d62-22ea3ab37ce5/volumes" Oct 03 16:52:48 crc kubenswrapper[4744]: I1003 16:52:48.029458 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-5696-account-create-jn6hh"] Oct 03 16:52:48 crc kubenswrapper[4744]: I1003 16:52:48.038044 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-5696-account-create-jn6hh"] Oct 03 16:52:48 crc kubenswrapper[4744]: I1003 16:52:48.912878 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d34f7f18-2333-4975-9606-9dd4ec282f3e" path="/var/lib/kubelet/pods/d34f7f18-2333-4975-9606-9dd4ec282f3e/volumes" Oct 03 16:52:50 crc kubenswrapper[4744]: I1003 16:52:50.892249 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:52:50 crc kubenswrapper[4744]: E1003 16:52:50.892923 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:53:05 crc kubenswrapper[4744]: I1003 16:53:05.892338 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:53:05 crc kubenswrapper[4744]: E1003 16:53:05.893211 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:53:09 crc kubenswrapper[4744]: I1003 16:53:09.054262 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-p5bkm"] Oct 03 16:53:09 crc kubenswrapper[4744]: I1003 16:53:09.066806 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-xrtkk"] Oct 03 16:53:09 crc kubenswrapper[4744]: I1003 16:53:09.077922 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-wdg7r"] Oct 03 16:53:09 crc kubenswrapper[4744]: I1003 16:53:09.086669 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-p5bkm"] Oct 03 16:53:09 crc kubenswrapper[4744]: I1003 16:53:09.095256 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-6j7qv"] Oct 03 16:53:09 crc kubenswrapper[4744]: I1003 16:53:09.103338 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-xrtkk"] Oct 03 16:53:09 crc kubenswrapper[4744]: I1003 16:53:09.111249 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-wdg7r"] Oct 03 16:53:09 crc kubenswrapper[4744]: I1003 16:53:09.117603 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-6j7qv"] Oct 03 16:53:10 crc kubenswrapper[4744]: I1003 16:53:10.906670 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="019c5b16-937c-45d2-9003-503f4d88995d" path="/var/lib/kubelet/pods/019c5b16-937c-45d2-9003-503f4d88995d/volumes" Oct 03 16:53:10 crc kubenswrapper[4744]: I1003 16:53:10.907640 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18c69bbd-4199-4da7-8e11-e05ea8680342" path="/var/lib/kubelet/pods/18c69bbd-4199-4da7-8e11-e05ea8680342/volumes" Oct 03 16:53:10 crc kubenswrapper[4744]: I1003 16:53:10.908211 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="299d80eb-fc43-4647-83be-a792412df7b1" path="/var/lib/kubelet/pods/299d80eb-fc43-4647-83be-a792412df7b1/volumes" Oct 03 16:53:10 crc kubenswrapper[4744]: I1003 16:53:10.908823 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97fcb340-8304-4bae-9d2b-f0b04dcea8bf" path="/var/lib/kubelet/pods/97fcb340-8304-4bae-9d2b-f0b04dcea8bf/volumes" Oct 03 16:53:12 crc kubenswrapper[4744]: I1003 16:53:12.041195 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-l78wg"] Oct 03 16:53:12 crc kubenswrapper[4744]: I1003 16:53:12.052057 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-l78wg"] Oct 03 16:53:12 crc kubenswrapper[4744]: I1003 16:53:12.913972 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d4a99fa-d668-4798-b988-a8d8fb9e3b28" path="/var/lib/kubelet/pods/3d4a99fa-d668-4798-b988-a8d8fb9e3b28/volumes" Oct 03 16:53:13 crc kubenswrapper[4744]: I1003 16:53:13.039594 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-cpdb5"] Oct 03 16:53:13 crc kubenswrapper[4744]: I1003 16:53:13.049338 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-cpdb5"] Oct 03 16:53:13 crc kubenswrapper[4744]: I1003 16:53:13.391874 4744 generic.go:334] "Generic (PLEG): container finished" podID="b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1" containerID="58b5309126059f9585226736dab9306fdc164b56b0e50872320b4d693fbc7009" exitCode=0 Oct 03 16:53:13 crc kubenswrapper[4744]: I1003 16:53:13.391896 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" event={"ID":"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1","Type":"ContainerDied","Data":"58b5309126059f9585226736dab9306fdc164b56b0e50872320b4d693fbc7009"} Oct 03 16:53:14 crc kubenswrapper[4744]: I1003 16:53:14.906786 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd5b11ee-b171-4bd8-9f40-34d9dcb428fe" path="/var/lib/kubelet/pods/cd5b11ee-b171-4bd8-9f40-34d9dcb428fe/volumes" Oct 03 16:53:14 crc kubenswrapper[4744]: I1003 16:53:14.972661 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.063540 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-ssh-key\") pod \"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1\" (UID: \"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1\") " Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.063835 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-inventory\") pod \"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1\" (UID: \"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1\") " Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.063877 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48zx5\" (UniqueName: \"kubernetes.io/projected/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-kube-api-access-48zx5\") pod \"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1\" (UID: \"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1\") " Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.070876 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-kube-api-access-48zx5" (OuterVolumeSpecName: "kube-api-access-48zx5") pod "b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1" (UID: "b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1"). InnerVolumeSpecName "kube-api-access-48zx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.104794 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1" (UID: "b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.130887 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-inventory" (OuterVolumeSpecName: "inventory") pod "b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1" (UID: "b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.166824 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.167070 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.167211 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48zx5\" (UniqueName: \"kubernetes.io/projected/b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1-kube-api-access-48zx5\") on node \"crc\" DevicePath \"\"" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.418380 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" event={"ID":"b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1","Type":"ContainerDied","Data":"01632ffddcb9cf44ce14db1738c5cbbe313048c43ecbe0366341f7cf54ebd341"} Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.418424 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01632ffddcb9cf44ce14db1738c5cbbe313048c43ecbe0366341f7cf54ebd341" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.418876 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.495583 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg"] Oct 03 16:53:15 crc kubenswrapper[4744]: E1003 16:53:15.495990 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.496010 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.496216 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.496901 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.499691 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.500427 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.500608 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.500822 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.506348 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg"] Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.676899 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg\" (UID: \"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.676963 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg\" (UID: \"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.677020 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqwbq\" (UniqueName: \"kubernetes.io/projected/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-kube-api-access-gqwbq\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg\" (UID: \"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.779122 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg\" (UID: \"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.779225 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg\" (UID: \"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.779294 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqwbq\" (UniqueName: \"kubernetes.io/projected/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-kube-api-access-gqwbq\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg\" (UID: \"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.784672 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg\" (UID: \"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.788778 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg\" (UID: \"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.796376 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqwbq\" (UniqueName: \"kubernetes.io/projected/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-kube-api-access-gqwbq\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg\" (UID: \"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" Oct 03 16:53:15 crc kubenswrapper[4744]: I1003 16:53:15.814628 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" Oct 03 16:53:16 crc kubenswrapper[4744]: I1003 16:53:16.441230 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg"] Oct 03 16:53:17 crc kubenswrapper[4744]: I1003 16:53:17.455075 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" event={"ID":"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce","Type":"ContainerStarted","Data":"0143c49c9020ea1eb2f01c716c8708e8ad82b5deb888ddff22cc1a12ff85d8b2"} Oct 03 16:53:17 crc kubenswrapper[4744]: I1003 16:53:17.455522 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" event={"ID":"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce","Type":"ContainerStarted","Data":"58551204b79f4bafa2fa79bf7cb6aab0784d08e1a042da3c469953af08f27bc0"} Oct 03 16:53:17 crc kubenswrapper[4744]: I1003 16:53:17.477902 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" podStartSLOduration=1.993733395 podStartE2EDuration="2.477878241s" podCreationTimestamp="2025-10-03 16:53:15 +0000 UTC" firstStartedPulling="2025-10-03 16:53:16.437150161 +0000 UTC m=+1722.717026057" lastFinishedPulling="2025-10-03 16:53:16.921295007 +0000 UTC m=+1723.201170903" observedRunningTime="2025-10-03 16:53:17.475954984 +0000 UTC m=+1723.755830920" watchObservedRunningTime="2025-10-03 16:53:17.477878241 +0000 UTC m=+1723.757754177" Oct 03 16:53:17 crc kubenswrapper[4744]: I1003 16:53:17.892102 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:53:17 crc kubenswrapper[4744]: E1003 16:53:17.893051 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:53:23 crc kubenswrapper[4744]: I1003 16:53:23.044991 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-2091-account-create-wz79h"] Oct 03 16:53:23 crc kubenswrapper[4744]: I1003 16:53:23.061960 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-2091-account-create-wz79h"] Oct 03 16:53:24 crc kubenswrapper[4744]: I1003 16:53:24.913190 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08780af0-0f6f-419c-8836-2e4661c5ff70" path="/var/lib/kubelet/pods/08780af0-0f6f-419c-8836-2e4661c5ff70/volumes" Oct 03 16:53:29 crc kubenswrapper[4744]: I1003 16:53:29.058194 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-0d09-account-create-g4bpt"] Oct 03 16:53:29 crc kubenswrapper[4744]: I1003 16:53:29.071105 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-0d09-account-create-g4bpt"] Oct 03 16:53:30 crc kubenswrapper[4744]: I1003 16:53:30.892347 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:53:30 crc kubenswrapper[4744]: E1003 16:53:30.893486 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:53:30 crc kubenswrapper[4744]: I1003 16:53:30.915604 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa76346f-0ca5-4bad-8fb8-802077202e39" path="/var/lib/kubelet/pods/aa76346f-0ca5-4bad-8fb8-802077202e39/volumes" Oct 03 16:53:31 crc kubenswrapper[4744]: I1003 16:53:31.034484 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-3d6b-account-create-klnll"] Oct 03 16:53:31 crc kubenswrapper[4744]: I1003 16:53:31.043604 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-3d6b-account-create-klnll"] Oct 03 16:53:31 crc kubenswrapper[4744]: I1003 16:53:31.052155 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-ffe1-account-create-nkjkf"] Oct 03 16:53:31 crc kubenswrapper[4744]: I1003 16:53:31.061205 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-ffe1-account-create-nkjkf"] Oct 03 16:53:32 crc kubenswrapper[4744]: I1003 16:53:32.907031 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9be7c905-97b5-40a2-b416-ff8894216fb4" path="/var/lib/kubelet/pods/9be7c905-97b5-40a2-b416-ff8894216fb4/volumes" Oct 03 16:53:32 crc kubenswrapper[4744]: I1003 16:53:32.908234 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4594c72-66b3-410f-b2ef-dbb205807338" path="/var/lib/kubelet/pods/c4594c72-66b3-410f-b2ef-dbb205807338/volumes" Oct 03 16:53:37 crc kubenswrapper[4744]: I1003 16:53:37.794447 4744 scope.go:117] "RemoveContainer" containerID="f505d3b8552d2d84f99a727f75a0601cc01798cdb6435537d32469da7debca61" Oct 03 16:53:37 crc kubenswrapper[4744]: I1003 16:53:37.861192 4744 scope.go:117] "RemoveContainer" containerID="6ed89bc00a03e9a367db4ca1449835b8957b051b02546da6ccde505e97e3630c" Oct 03 16:53:37 crc kubenswrapper[4744]: I1003 16:53:37.921751 4744 scope.go:117] "RemoveContainer" containerID="0991e3d7331305f7c44f8d526bf5753a3739f1d50ac22a91c69869006166c4c1" Oct 03 16:53:37 crc kubenswrapper[4744]: I1003 16:53:37.973886 4744 scope.go:117] "RemoveContainer" containerID="d9506b1991e30e193a66b300a91f395b1c2458fa895fdd1f8798c4753a65a1ba" Oct 03 16:53:38 crc kubenswrapper[4744]: I1003 16:53:38.013647 4744 scope.go:117] "RemoveContainer" containerID="c9f5c9f33347190f940722acea6862eb028c2c50d98422eac5885f8ea9f40616" Oct 03 16:53:38 crc kubenswrapper[4744]: I1003 16:53:38.064215 4744 scope.go:117] "RemoveContainer" containerID="3a939c13f0ada6e3771f4b1177d843e7975333c60e7bd85ce7a7a3e24e5ab528" Oct 03 16:53:38 crc kubenswrapper[4744]: I1003 16:53:38.103114 4744 scope.go:117] "RemoveContainer" containerID="aa607b01ba85a1bd8828075e6bd1127d39a8420d2c70e3da63381ffd39d72f89" Oct 03 16:53:38 crc kubenswrapper[4744]: I1003 16:53:38.130137 4744 scope.go:117] "RemoveContainer" containerID="798f39b84e71fc24d78389f8bd3f1c41309abc68525621cb19513b377551f6e9" Oct 03 16:53:38 crc kubenswrapper[4744]: I1003 16:53:38.164461 4744 scope.go:117] "RemoveContainer" containerID="04dc55f54d87503af16952d67d9653f221b6abe42f04b82718de2e1d235511f5" Oct 03 16:53:38 crc kubenswrapper[4744]: I1003 16:53:38.200835 4744 scope.go:117] "RemoveContainer" containerID="786ebc5ff90c568b00f25cde96f5779e594932c2c57b36ba551e611472ffa602" Oct 03 16:53:38 crc kubenswrapper[4744]: I1003 16:53:38.228176 4744 scope.go:117] "RemoveContainer" containerID="40905c154d75a80574ac970ee03bdfa5e8f403139e5cb5aef9cc275934ef0411" Oct 03 16:53:38 crc kubenswrapper[4744]: I1003 16:53:38.247635 4744 scope.go:117] "RemoveContainer" containerID="7b951a5e87e97262961e2d9245e35801cb16b1e333e328900801c6e43581e69b" Oct 03 16:53:41 crc kubenswrapper[4744]: I1003 16:53:41.894242 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:53:41 crc kubenswrapper[4744]: E1003 16:53:41.900012 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:53:47 crc kubenswrapper[4744]: I1003 16:53:47.050390 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-9ftfh"] Oct 03 16:53:47 crc kubenswrapper[4744]: I1003 16:53:47.062648 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-9ftfh"] Oct 03 16:53:48 crc kubenswrapper[4744]: I1003 16:53:48.908775 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="245dc938-9219-43e5-8e69-10e033862451" path="/var/lib/kubelet/pods/245dc938-9219-43e5-8e69-10e033862451/volumes" Oct 03 16:53:54 crc kubenswrapper[4744]: I1003 16:53:54.908599 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:53:54 crc kubenswrapper[4744]: E1003 16:53:54.909628 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:54:02 crc kubenswrapper[4744]: I1003 16:54:02.044467 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-b88bn"] Oct 03 16:54:02 crc kubenswrapper[4744]: I1003 16:54:02.053788 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-b88bn"] Oct 03 16:54:02 crc kubenswrapper[4744]: I1003 16:54:02.906055 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7" path="/var/lib/kubelet/pods/3e2e9c95-2d1a-495f-9d3f-7433fd43dcb7/volumes" Oct 03 16:54:07 crc kubenswrapper[4744]: I1003 16:54:07.892694 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:54:07 crc kubenswrapper[4744]: E1003 16:54:07.893568 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:54:17 crc kubenswrapper[4744]: I1003 16:54:17.031663 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-qbz46"] Oct 03 16:54:17 crc kubenswrapper[4744]: I1003 16:54:17.042452 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-qbz46"] Oct 03 16:54:18 crc kubenswrapper[4744]: I1003 16:54:18.896470 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:54:18 crc kubenswrapper[4744]: E1003 16:54:18.898227 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:54:18 crc kubenswrapper[4744]: I1003 16:54:18.913871 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7acfc27b-28b7-422c-a933-8dd2a4a99c68" path="/var/lib/kubelet/pods/7acfc27b-28b7-422c-a933-8dd2a4a99c68/volumes" Oct 03 16:54:21 crc kubenswrapper[4744]: I1003 16:54:21.238918 4744 generic.go:334] "Generic (PLEG): container finished" podID="4fe42e89-74c8-4ea1-b1f5-af28a032f3ce" containerID="0143c49c9020ea1eb2f01c716c8708e8ad82b5deb888ddff22cc1a12ff85d8b2" exitCode=0 Oct 03 16:54:21 crc kubenswrapper[4744]: I1003 16:54:21.239047 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" event={"ID":"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce","Type":"ContainerDied","Data":"0143c49c9020ea1eb2f01c716c8708e8ad82b5deb888ddff22cc1a12ff85d8b2"} Oct 03 16:54:22 crc kubenswrapper[4744]: I1003 16:54:22.735190 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" Oct 03 16:54:22 crc kubenswrapper[4744]: I1003 16:54:22.911509 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-inventory\") pod \"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce\" (UID: \"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce\") " Oct 03 16:54:22 crc kubenswrapper[4744]: I1003 16:54:22.911729 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-ssh-key\") pod \"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce\" (UID: \"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce\") " Oct 03 16:54:22 crc kubenswrapper[4744]: I1003 16:54:22.911813 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqwbq\" (UniqueName: \"kubernetes.io/projected/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-kube-api-access-gqwbq\") pod \"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce\" (UID: \"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce\") " Oct 03 16:54:22 crc kubenswrapper[4744]: I1003 16:54:22.919815 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-kube-api-access-gqwbq" (OuterVolumeSpecName: "kube-api-access-gqwbq") pod "4fe42e89-74c8-4ea1-b1f5-af28a032f3ce" (UID: "4fe42e89-74c8-4ea1-b1f5-af28a032f3ce"). InnerVolumeSpecName "kube-api-access-gqwbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:54:22 crc kubenswrapper[4744]: I1003 16:54:22.948600 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4fe42e89-74c8-4ea1-b1f5-af28a032f3ce" (UID: "4fe42e89-74c8-4ea1-b1f5-af28a032f3ce"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:54:22 crc kubenswrapper[4744]: I1003 16:54:22.951052 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-inventory" (OuterVolumeSpecName: "inventory") pod "4fe42e89-74c8-4ea1-b1f5-af28a032f3ce" (UID: "4fe42e89-74c8-4ea1-b1f5-af28a032f3ce"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.017037 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.017855 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.017962 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqwbq\" (UniqueName: \"kubernetes.io/projected/4fe42e89-74c8-4ea1-b1f5-af28a032f3ce-kube-api-access-gqwbq\") on node \"crc\" DevicePath \"\"" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.028835 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-6k6b8"] Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.037608 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-6k6b8"] Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.268167 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" event={"ID":"4fe42e89-74c8-4ea1-b1f5-af28a032f3ce","Type":"ContainerDied","Data":"58551204b79f4bafa2fa79bf7cb6aab0784d08e1a042da3c469953af08f27bc0"} Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.268572 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58551204b79f4bafa2fa79bf7cb6aab0784d08e1a042da3c469953af08f27bc0" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.268275 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.346223 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7"] Oct 03 16:54:23 crc kubenswrapper[4744]: E1003 16:54:23.346642 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fe42e89-74c8-4ea1-b1f5-af28a032f3ce" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.346663 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fe42e89-74c8-4ea1-b1f5-af28a032f3ce" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.346851 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fe42e89-74c8-4ea1-b1f5-af28a032f3ce" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.347441 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.350626 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.350636 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.350636 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.350663 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.368860 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7"] Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.424107 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmf2z\" (UniqueName: \"kubernetes.io/projected/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-kube-api-access-qmf2z\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7\" (UID: \"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.424192 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7\" (UID: \"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.424364 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7\" (UID: \"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.526321 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmf2z\" (UniqueName: \"kubernetes.io/projected/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-kube-api-access-qmf2z\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7\" (UID: \"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.526410 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7\" (UID: \"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.526510 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7\" (UID: \"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.531932 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7\" (UID: \"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.533945 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7\" (UID: \"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.562347 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmf2z\" (UniqueName: \"kubernetes.io/projected/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-kube-api-access-qmf2z\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7\" (UID: \"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" Oct 03 16:54:23 crc kubenswrapper[4744]: I1003 16:54:23.672929 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" Oct 03 16:54:24 crc kubenswrapper[4744]: I1003 16:54:24.036701 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7"] Oct 03 16:54:24 crc kubenswrapper[4744]: I1003 16:54:24.281190 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" event={"ID":"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b","Type":"ContainerStarted","Data":"dfcc25b84fb0b87e816a08bc83efd474bb57387d240e8c6d9d01368ed214836e"} Oct 03 16:54:24 crc kubenswrapper[4744]: I1003 16:54:24.904637 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc8def8d-6149-4797-b453-103ec15579f6" path="/var/lib/kubelet/pods/dc8def8d-6149-4797-b453-103ec15579f6/volumes" Oct 03 16:54:25 crc kubenswrapper[4744]: I1003 16:54:25.291472 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" event={"ID":"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b","Type":"ContainerStarted","Data":"bbfa6578b93d58e3925622962fcba4cb581decbee09eb03ebe36bfdac5584458"} Oct 03 16:54:25 crc kubenswrapper[4744]: I1003 16:54:25.313718 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" podStartSLOduration=1.666281429 podStartE2EDuration="2.313698966s" podCreationTimestamp="2025-10-03 16:54:23 +0000 UTC" firstStartedPulling="2025-10-03 16:54:24.042176035 +0000 UTC m=+1790.322051931" lastFinishedPulling="2025-10-03 16:54:24.689593552 +0000 UTC m=+1790.969469468" observedRunningTime="2025-10-03 16:54:25.312523557 +0000 UTC m=+1791.592399483" watchObservedRunningTime="2025-10-03 16:54:25.313698966 +0000 UTC m=+1791.593574872" Oct 03 16:54:28 crc kubenswrapper[4744]: I1003 16:54:28.074933 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-krczc"] Oct 03 16:54:28 crc kubenswrapper[4744]: I1003 16:54:28.089881 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-krczc"] Oct 03 16:54:28 crc kubenswrapper[4744]: I1003 16:54:28.907173 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8054599-30d8-42e0-bf45-2a8758b10486" path="/var/lib/kubelet/pods/a8054599-30d8-42e0-bf45-2a8758b10486/volumes" Oct 03 16:54:29 crc kubenswrapper[4744]: I1003 16:54:29.892330 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:54:29 crc kubenswrapper[4744]: E1003 16:54:29.893008 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:54:30 crc kubenswrapper[4744]: I1003 16:54:30.336972 4744 generic.go:334] "Generic (PLEG): container finished" podID="d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b" containerID="bbfa6578b93d58e3925622962fcba4cb581decbee09eb03ebe36bfdac5584458" exitCode=0 Oct 03 16:54:30 crc kubenswrapper[4744]: I1003 16:54:30.337085 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" event={"ID":"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b","Type":"ContainerDied","Data":"bbfa6578b93d58e3925622962fcba4cb581decbee09eb03ebe36bfdac5584458"} Oct 03 16:54:31 crc kubenswrapper[4744]: I1003 16:54:31.760587 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" Oct 03 16:54:31 crc kubenswrapper[4744]: I1003 16:54:31.821045 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmf2z\" (UniqueName: \"kubernetes.io/projected/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-kube-api-access-qmf2z\") pod \"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b\" (UID: \"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b\") " Oct 03 16:54:31 crc kubenswrapper[4744]: I1003 16:54:31.821168 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-inventory\") pod \"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b\" (UID: \"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b\") " Oct 03 16:54:31 crc kubenswrapper[4744]: I1003 16:54:31.821275 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-ssh-key\") pod \"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b\" (UID: \"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b\") " Oct 03 16:54:31 crc kubenswrapper[4744]: I1003 16:54:31.828913 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-kube-api-access-qmf2z" (OuterVolumeSpecName: "kube-api-access-qmf2z") pod "d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b" (UID: "d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b"). InnerVolumeSpecName "kube-api-access-qmf2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:54:31 crc kubenswrapper[4744]: I1003 16:54:31.851897 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-inventory" (OuterVolumeSpecName: "inventory") pod "d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b" (UID: "d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:54:31 crc kubenswrapper[4744]: I1003 16:54:31.857976 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b" (UID: "d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:54:31 crc kubenswrapper[4744]: I1003 16:54:31.923096 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:54:31 crc kubenswrapper[4744]: I1003 16:54:31.923133 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmf2z\" (UniqueName: \"kubernetes.io/projected/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-kube-api-access-qmf2z\") on node \"crc\" DevicePath \"\"" Oct 03 16:54:31 crc kubenswrapper[4744]: I1003 16:54:31.923254 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.354868 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" event={"ID":"d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b","Type":"ContainerDied","Data":"dfcc25b84fb0b87e816a08bc83efd474bb57387d240e8c6d9d01368ed214836e"} Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.355118 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfcc25b84fb0b87e816a08bc83efd474bb57387d240e8c6d9d01368ed214836e" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.354929 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.430954 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68"] Oct 03 16:54:32 crc kubenswrapper[4744]: E1003 16:54:32.431462 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.431486 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.431865 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.432800 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.435065 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.435075 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.435590 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.436213 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.441948 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68"] Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.535199 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntjr5\" (UniqueName: \"kubernetes.io/projected/42a1ea96-7b56-4594-811d-7043e8640e39-kube-api-access-ntjr5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jbg68\" (UID: \"42a1ea96-7b56-4594-811d-7043e8640e39\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.535241 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42a1ea96-7b56-4594-811d-7043e8640e39-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jbg68\" (UID: \"42a1ea96-7b56-4594-811d-7043e8640e39\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.535350 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42a1ea96-7b56-4594-811d-7043e8640e39-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jbg68\" (UID: \"42a1ea96-7b56-4594-811d-7043e8640e39\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.637971 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42a1ea96-7b56-4594-811d-7043e8640e39-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jbg68\" (UID: \"42a1ea96-7b56-4594-811d-7043e8640e39\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.638228 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntjr5\" (UniqueName: \"kubernetes.io/projected/42a1ea96-7b56-4594-811d-7043e8640e39-kube-api-access-ntjr5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jbg68\" (UID: \"42a1ea96-7b56-4594-811d-7043e8640e39\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.638280 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42a1ea96-7b56-4594-811d-7043e8640e39-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jbg68\" (UID: \"42a1ea96-7b56-4594-811d-7043e8640e39\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.641786 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42a1ea96-7b56-4594-811d-7043e8640e39-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jbg68\" (UID: \"42a1ea96-7b56-4594-811d-7043e8640e39\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.644720 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42a1ea96-7b56-4594-811d-7043e8640e39-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jbg68\" (UID: \"42a1ea96-7b56-4594-811d-7043e8640e39\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.664104 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntjr5\" (UniqueName: \"kubernetes.io/projected/42a1ea96-7b56-4594-811d-7043e8640e39-kube-api-access-ntjr5\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jbg68\" (UID: \"42a1ea96-7b56-4594-811d-7043e8640e39\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" Oct 03 16:54:32 crc kubenswrapper[4744]: I1003 16:54:32.768003 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" Oct 03 16:54:33 crc kubenswrapper[4744]: I1003 16:54:33.357417 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68"] Oct 03 16:54:33 crc kubenswrapper[4744]: I1003 16:54:33.364899 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" event={"ID":"42a1ea96-7b56-4594-811d-7043e8640e39","Type":"ContainerStarted","Data":"5ca89c8a1071efa4377ed16d30a323a45dfc40639cfbf387a43b6d126cd4863a"} Oct 03 16:54:35 crc kubenswrapper[4744]: I1003 16:54:35.394134 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" event={"ID":"42a1ea96-7b56-4594-811d-7043e8640e39","Type":"ContainerStarted","Data":"8f9a941e0cbdaad07915a7018cb0292f1a0ab5aadfac1f1dbaf8d074199dfe84"} Oct 03 16:54:35 crc kubenswrapper[4744]: I1003 16:54:35.427601 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" podStartSLOduration=2.591837864 podStartE2EDuration="3.427580019s" podCreationTimestamp="2025-10-03 16:54:32 +0000 UTC" firstStartedPulling="2025-10-03 16:54:33.356917621 +0000 UTC m=+1799.636793517" lastFinishedPulling="2025-10-03 16:54:34.192659766 +0000 UTC m=+1800.472535672" observedRunningTime="2025-10-03 16:54:35.423882148 +0000 UTC m=+1801.703758054" watchObservedRunningTime="2025-10-03 16:54:35.427580019 +0000 UTC m=+1801.707455915" Oct 03 16:54:38 crc kubenswrapper[4744]: I1003 16:54:38.470671 4744 scope.go:117] "RemoveContainer" containerID="0802af45ac185ed4e7848dce58259daaefb0b25cf43bdc9fae87eebd75be34ed" Oct 03 16:54:38 crc kubenswrapper[4744]: I1003 16:54:38.522616 4744 scope.go:117] "RemoveContainer" containerID="8c2b637e9fea89323ced9b6a7c16ff4a1499d858e2fb5fcb8ba4fb3b95f37182" Oct 03 16:54:38 crc kubenswrapper[4744]: I1003 16:54:38.551063 4744 scope.go:117] "RemoveContainer" containerID="0fe5a6ae0fcdb499668ae953e15d63f01addf862e36032e8e66e70b365a4ba0c" Oct 03 16:54:38 crc kubenswrapper[4744]: I1003 16:54:38.596632 4744 scope.go:117] "RemoveContainer" containerID="6286e9ec39d254af81ef9ab41125f9cdbd2ce065a205572c2bbe1eb4f3a4bd0f" Oct 03 16:54:38 crc kubenswrapper[4744]: I1003 16:54:38.658920 4744 scope.go:117] "RemoveContainer" containerID="7aa00cc7e148d70ea0f891901929ff502cafa279aba0996289a3c2ae059e648a" Oct 03 16:54:41 crc kubenswrapper[4744]: I1003 16:54:41.892532 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:54:41 crc kubenswrapper[4744]: E1003 16:54:41.893830 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:54:42 crc kubenswrapper[4744]: I1003 16:54:42.050652 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-8bj8c"] Oct 03 16:54:42 crc kubenswrapper[4744]: I1003 16:54:42.065825 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-8bj8c"] Oct 03 16:54:42 crc kubenswrapper[4744]: I1003 16:54:42.911727 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df7dff0d-99e0-4e98-b752-9096d24e314b" path="/var/lib/kubelet/pods/df7dff0d-99e0-4e98-b752-9096d24e314b/volumes" Oct 03 16:54:54 crc kubenswrapper[4744]: I1003 16:54:54.902386 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:54:54 crc kubenswrapper[4744]: E1003 16:54:54.903463 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:55:02 crc kubenswrapper[4744]: I1003 16:55:02.056426 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-wh96r"] Oct 03 16:55:02 crc kubenswrapper[4744]: I1003 16:55:02.066559 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-wh96r"] Oct 03 16:55:02 crc kubenswrapper[4744]: I1003 16:55:02.911300 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cca8b080-5315-4c0e-ab9d-c94c248df68d" path="/var/lib/kubelet/pods/cca8b080-5315-4c0e-ab9d-c94c248df68d/volumes" Oct 03 16:55:03 crc kubenswrapper[4744]: I1003 16:55:03.046602 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-ldg2v"] Oct 03 16:55:03 crc kubenswrapper[4744]: I1003 16:55:03.053682 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-qsplt"] Oct 03 16:55:03 crc kubenswrapper[4744]: I1003 16:55:03.062175 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-ldg2v"] Oct 03 16:55:03 crc kubenswrapper[4744]: I1003 16:55:03.070559 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-qsplt"] Oct 03 16:55:04 crc kubenswrapper[4744]: I1003 16:55:04.908565 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15b5b2d9-3e50-4326-a345-8b5f79d684e2" path="/var/lib/kubelet/pods/15b5b2d9-3e50-4326-a345-8b5f79d684e2/volumes" Oct 03 16:55:04 crc kubenswrapper[4744]: I1003 16:55:04.909582 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="501a4d70-6aa9-4fd5-8e72-5b854e648219" path="/var/lib/kubelet/pods/501a4d70-6aa9-4fd5-8e72-5b854e648219/volumes" Oct 03 16:55:06 crc kubenswrapper[4744]: I1003 16:55:06.892613 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:55:06 crc kubenswrapper[4744]: E1003 16:55:06.893268 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:55:09 crc kubenswrapper[4744]: I1003 16:55:09.044779 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-b978-account-create-4wzh7"] Oct 03 16:55:09 crc kubenswrapper[4744]: I1003 16:55:09.061516 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-b978-account-create-4wzh7"] Oct 03 16:55:10 crc kubenswrapper[4744]: I1003 16:55:10.913004 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9826741-7031-4894-8ff1-ad1a150f367a" path="/var/lib/kubelet/pods/e9826741-7031-4894-8ff1-ad1a150f367a/volumes" Oct 03 16:55:11 crc kubenswrapper[4744]: E1003 16:55:11.445668 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42a1ea96_7b56_4594_811d_7043e8640e39.slice/crio-conmon-8f9a941e0cbdaad07915a7018cb0292f1a0ab5aadfac1f1dbaf8d074199dfe84.scope\": RecentStats: unable to find data in memory cache]" Oct 03 16:55:11 crc kubenswrapper[4744]: I1003 16:55:11.796307 4744 generic.go:334] "Generic (PLEG): container finished" podID="42a1ea96-7b56-4594-811d-7043e8640e39" containerID="8f9a941e0cbdaad07915a7018cb0292f1a0ab5aadfac1f1dbaf8d074199dfe84" exitCode=0 Oct 03 16:55:11 crc kubenswrapper[4744]: I1003 16:55:11.796357 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" event={"ID":"42a1ea96-7b56-4594-811d-7043e8640e39","Type":"ContainerDied","Data":"8f9a941e0cbdaad07915a7018cb0292f1a0ab5aadfac1f1dbaf8d074199dfe84"} Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.347620 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.367582 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42a1ea96-7b56-4594-811d-7043e8640e39-inventory\") pod \"42a1ea96-7b56-4594-811d-7043e8640e39\" (UID: \"42a1ea96-7b56-4594-811d-7043e8640e39\") " Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.367690 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42a1ea96-7b56-4594-811d-7043e8640e39-ssh-key\") pod \"42a1ea96-7b56-4594-811d-7043e8640e39\" (UID: \"42a1ea96-7b56-4594-811d-7043e8640e39\") " Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.367983 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntjr5\" (UniqueName: \"kubernetes.io/projected/42a1ea96-7b56-4594-811d-7043e8640e39-kube-api-access-ntjr5\") pod \"42a1ea96-7b56-4594-811d-7043e8640e39\" (UID: \"42a1ea96-7b56-4594-811d-7043e8640e39\") " Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.377265 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42a1ea96-7b56-4594-811d-7043e8640e39-kube-api-access-ntjr5" (OuterVolumeSpecName: "kube-api-access-ntjr5") pod "42a1ea96-7b56-4594-811d-7043e8640e39" (UID: "42a1ea96-7b56-4594-811d-7043e8640e39"). InnerVolumeSpecName "kube-api-access-ntjr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.421692 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42a1ea96-7b56-4594-811d-7043e8640e39-inventory" (OuterVolumeSpecName: "inventory") pod "42a1ea96-7b56-4594-811d-7043e8640e39" (UID: "42a1ea96-7b56-4594-811d-7043e8640e39"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.426942 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42a1ea96-7b56-4594-811d-7043e8640e39-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "42a1ea96-7b56-4594-811d-7043e8640e39" (UID: "42a1ea96-7b56-4594-811d-7043e8640e39"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.471256 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntjr5\" (UniqueName: \"kubernetes.io/projected/42a1ea96-7b56-4594-811d-7043e8640e39-kube-api-access-ntjr5\") on node \"crc\" DevicePath \"\"" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.471291 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42a1ea96-7b56-4594-811d-7043e8640e39-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.471301 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/42a1ea96-7b56-4594-811d-7043e8640e39-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.826633 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" event={"ID":"42a1ea96-7b56-4594-811d-7043e8640e39","Type":"ContainerDied","Data":"5ca89c8a1071efa4377ed16d30a323a45dfc40639cfbf387a43b6d126cd4863a"} Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.826697 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jbg68" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.826710 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ca89c8a1071efa4377ed16d30a323a45dfc40639cfbf387a43b6d126cd4863a" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.983411 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz"] Oct 03 16:55:13 crc kubenswrapper[4744]: E1003 16:55:13.983986 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42a1ea96-7b56-4594-811d-7043e8640e39" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.984018 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="42a1ea96-7b56-4594-811d-7043e8640e39" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.984303 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="42a1ea96-7b56-4594-811d-7043e8640e39" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.985341 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.986977 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.988092 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.988373 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 16:55:13 crc kubenswrapper[4744]: I1003 16:55:13.988389 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 16:55:14 crc kubenswrapper[4744]: I1003 16:55:14.005035 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz"] Oct 03 16:55:14 crc kubenswrapper[4744]: I1003 16:55:14.084863 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af14e991-2a54-4228-aa49-a8ce99d8eab2-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz\" (UID: \"af14e991-2a54-4228-aa49-a8ce99d8eab2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" Oct 03 16:55:14 crc kubenswrapper[4744]: I1003 16:55:14.085037 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af14e991-2a54-4228-aa49-a8ce99d8eab2-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz\" (UID: \"af14e991-2a54-4228-aa49-a8ce99d8eab2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" Oct 03 16:55:14 crc kubenswrapper[4744]: I1003 16:55:14.085099 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ggwp\" (UniqueName: \"kubernetes.io/projected/af14e991-2a54-4228-aa49-a8ce99d8eab2-kube-api-access-8ggwp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz\" (UID: \"af14e991-2a54-4228-aa49-a8ce99d8eab2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" Oct 03 16:55:14 crc kubenswrapper[4744]: I1003 16:55:14.187882 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af14e991-2a54-4228-aa49-a8ce99d8eab2-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz\" (UID: \"af14e991-2a54-4228-aa49-a8ce99d8eab2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" Oct 03 16:55:14 crc kubenswrapper[4744]: I1003 16:55:14.188116 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af14e991-2a54-4228-aa49-a8ce99d8eab2-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz\" (UID: \"af14e991-2a54-4228-aa49-a8ce99d8eab2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" Oct 03 16:55:14 crc kubenswrapper[4744]: I1003 16:55:14.188170 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ggwp\" (UniqueName: \"kubernetes.io/projected/af14e991-2a54-4228-aa49-a8ce99d8eab2-kube-api-access-8ggwp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz\" (UID: \"af14e991-2a54-4228-aa49-a8ce99d8eab2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" Oct 03 16:55:14 crc kubenswrapper[4744]: I1003 16:55:14.192615 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af14e991-2a54-4228-aa49-a8ce99d8eab2-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz\" (UID: \"af14e991-2a54-4228-aa49-a8ce99d8eab2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" Oct 03 16:55:14 crc kubenswrapper[4744]: I1003 16:55:14.196247 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af14e991-2a54-4228-aa49-a8ce99d8eab2-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz\" (UID: \"af14e991-2a54-4228-aa49-a8ce99d8eab2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" Oct 03 16:55:14 crc kubenswrapper[4744]: I1003 16:55:14.212960 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ggwp\" (UniqueName: \"kubernetes.io/projected/af14e991-2a54-4228-aa49-a8ce99d8eab2-kube-api-access-8ggwp\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz\" (UID: \"af14e991-2a54-4228-aa49-a8ce99d8eab2\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" Oct 03 16:55:14 crc kubenswrapper[4744]: I1003 16:55:14.305278 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" Oct 03 16:55:14 crc kubenswrapper[4744]: I1003 16:55:14.877133 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz"] Oct 03 16:55:15 crc kubenswrapper[4744]: I1003 16:55:15.848596 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" event={"ID":"af14e991-2a54-4228-aa49-a8ce99d8eab2","Type":"ContainerStarted","Data":"0e0742204bc81a078106cc376aba23d5b3ca8d8e96c75974064d4c6dfa22cfc8"} Oct 03 16:55:15 crc kubenswrapper[4744]: I1003 16:55:15.848976 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" event={"ID":"af14e991-2a54-4228-aa49-a8ce99d8eab2","Type":"ContainerStarted","Data":"22a7650ab93b617444293b706c863fd854b16c3037e27f691d0318fd762257ea"} Oct 03 16:55:19 crc kubenswrapper[4744]: I1003 16:55:19.037072 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" podStartSLOduration=5.494935683 podStartE2EDuration="6.037039692s" podCreationTimestamp="2025-10-03 16:55:13 +0000 UTC" firstStartedPulling="2025-10-03 16:55:14.886167584 +0000 UTC m=+1841.166043480" lastFinishedPulling="2025-10-03 16:55:15.428271573 +0000 UTC m=+1841.708147489" observedRunningTime="2025-10-03 16:55:15.877369298 +0000 UTC m=+1842.157245254" watchObservedRunningTime="2025-10-03 16:55:19.037039692 +0000 UTC m=+1845.316915618" Oct 03 16:55:19 crc kubenswrapper[4744]: I1003 16:55:19.049064 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-fc01-account-create-5cmg6"] Oct 03 16:55:19 crc kubenswrapper[4744]: I1003 16:55:19.064861 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-7097-account-create-kzss9"] Oct 03 16:55:19 crc kubenswrapper[4744]: I1003 16:55:19.074420 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-7097-account-create-kzss9"] Oct 03 16:55:19 crc kubenswrapper[4744]: I1003 16:55:19.083580 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-fc01-account-create-5cmg6"] Oct 03 16:55:19 crc kubenswrapper[4744]: I1003 16:55:19.892025 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:55:19 crc kubenswrapper[4744]: E1003 16:55:19.892891 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:55:20 crc kubenswrapper[4744]: I1003 16:55:20.911955 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beff32b5-71c6-4be0-8a7f-5abae594db52" path="/var/lib/kubelet/pods/beff32b5-71c6-4be0-8a7f-5abae594db52/volumes" Oct 03 16:55:20 crc kubenswrapper[4744]: I1003 16:55:20.912865 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5" path="/var/lib/kubelet/pods/fe920b2a-f603-49a7-a0a6-c4e8ce9ff4f5/volumes" Oct 03 16:55:31 crc kubenswrapper[4744]: I1003 16:55:31.892983 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:55:31 crc kubenswrapper[4744]: E1003 16:55:31.893858 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:55:38 crc kubenswrapper[4744]: I1003 16:55:38.835393 4744 scope.go:117] "RemoveContainer" containerID="0ea1cd99b96b79967a4515714aea8d00a35506053339e368c6622b035958e0d4" Oct 03 16:55:38 crc kubenswrapper[4744]: I1003 16:55:38.871874 4744 scope.go:117] "RemoveContainer" containerID="bafe7e2bfc3a5219bf6833e1abe232b87d2682a3d9b40fadd8034609386721da" Oct 03 16:55:38 crc kubenswrapper[4744]: I1003 16:55:38.947440 4744 scope.go:117] "RemoveContainer" containerID="af81ada8db847b37ff88c6b34fd623097cd5f8b6bd7f1cac3c532aa7ca4eb1a3" Oct 03 16:55:39 crc kubenswrapper[4744]: I1003 16:55:39.009228 4744 scope.go:117] "RemoveContainer" containerID="9a137d288bf9afcc935660a6c8cc4377050731d231f1dbb19e8a65660eaf297b" Oct 03 16:55:39 crc kubenswrapper[4744]: I1003 16:55:39.060417 4744 scope.go:117] "RemoveContainer" containerID="f5a1edcf8dc2bd8f953414f1291177ce5631944d6d835b74686c4d7230d511fa" Oct 03 16:55:39 crc kubenswrapper[4744]: I1003 16:55:39.092680 4744 scope.go:117] "RemoveContainer" containerID="e7656f6bbb2b5a8116c6daa66427fab681e8be77ee5904e4f595b06426e76e0e" Oct 03 16:55:39 crc kubenswrapper[4744]: I1003 16:55:39.142005 4744 scope.go:117] "RemoveContainer" containerID="259b7d4a70bf2684be3f6247908706dd5d345163f48a7840387c842a166e3fb2" Oct 03 16:55:44 crc kubenswrapper[4744]: I1003 16:55:44.059851 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zgrsm"] Oct 03 16:55:44 crc kubenswrapper[4744]: I1003 16:55:44.075844 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-zgrsm"] Oct 03 16:55:44 crc kubenswrapper[4744]: I1003 16:55:44.908121 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cd09fa9-4967-49a1-a086-55405a0a681b" path="/var/lib/kubelet/pods/1cd09fa9-4967-49a1-a086-55405a0a681b/volumes" Oct 03 16:55:46 crc kubenswrapper[4744]: I1003 16:55:46.892748 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:55:46 crc kubenswrapper[4744]: E1003 16:55:46.893615 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:55:57 crc kubenswrapper[4744]: I1003 16:55:57.892721 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:55:57 crc kubenswrapper[4744]: E1003 16:55:57.893908 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:56:05 crc kubenswrapper[4744]: I1003 16:56:05.097968 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-g52js"] Oct 03 16:56:05 crc kubenswrapper[4744]: I1003 16:56:05.113762 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-g52js"] Oct 03 16:56:06 crc kubenswrapper[4744]: I1003 16:56:06.037428 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qhgd6"] Oct 03 16:56:06 crc kubenswrapper[4744]: I1003 16:56:06.048982 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qhgd6"] Oct 03 16:56:06 crc kubenswrapper[4744]: I1003 16:56:06.907214 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4faff3af-3870-4a9b-9fcf-91f0ecc30a8b" path="/var/lib/kubelet/pods/4faff3af-3870-4a9b-9fcf-91f0ecc30a8b/volumes" Oct 03 16:56:06 crc kubenswrapper[4744]: I1003 16:56:06.908478 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb1cb635-71ed-4d6f-89ba-b2dd9342a55a" path="/var/lib/kubelet/pods/bb1cb635-71ed-4d6f-89ba-b2dd9342a55a/volumes" Oct 03 16:56:09 crc kubenswrapper[4744]: I1003 16:56:09.591146 4744 generic.go:334] "Generic (PLEG): container finished" podID="af14e991-2a54-4228-aa49-a8ce99d8eab2" containerID="0e0742204bc81a078106cc376aba23d5b3ca8d8e96c75974064d4c6dfa22cfc8" exitCode=2 Oct 03 16:56:09 crc kubenswrapper[4744]: I1003 16:56:09.592044 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" event={"ID":"af14e991-2a54-4228-aa49-a8ce99d8eab2","Type":"ContainerDied","Data":"0e0742204bc81a078106cc376aba23d5b3ca8d8e96c75974064d4c6dfa22cfc8"} Oct 03 16:56:10 crc kubenswrapper[4744]: I1003 16:56:10.901878 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:56:10 crc kubenswrapper[4744]: E1003 16:56:10.902576 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:56:11 crc kubenswrapper[4744]: I1003 16:56:11.067640 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" Oct 03 16:56:11 crc kubenswrapper[4744]: I1003 16:56:11.257349 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af14e991-2a54-4228-aa49-a8ce99d8eab2-ssh-key\") pod \"af14e991-2a54-4228-aa49-a8ce99d8eab2\" (UID: \"af14e991-2a54-4228-aa49-a8ce99d8eab2\") " Oct 03 16:56:11 crc kubenswrapper[4744]: I1003 16:56:11.257441 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ggwp\" (UniqueName: \"kubernetes.io/projected/af14e991-2a54-4228-aa49-a8ce99d8eab2-kube-api-access-8ggwp\") pod \"af14e991-2a54-4228-aa49-a8ce99d8eab2\" (UID: \"af14e991-2a54-4228-aa49-a8ce99d8eab2\") " Oct 03 16:56:11 crc kubenswrapper[4744]: I1003 16:56:11.257474 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af14e991-2a54-4228-aa49-a8ce99d8eab2-inventory\") pod \"af14e991-2a54-4228-aa49-a8ce99d8eab2\" (UID: \"af14e991-2a54-4228-aa49-a8ce99d8eab2\") " Oct 03 16:56:11 crc kubenswrapper[4744]: I1003 16:56:11.268292 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af14e991-2a54-4228-aa49-a8ce99d8eab2-kube-api-access-8ggwp" (OuterVolumeSpecName: "kube-api-access-8ggwp") pod "af14e991-2a54-4228-aa49-a8ce99d8eab2" (UID: "af14e991-2a54-4228-aa49-a8ce99d8eab2"). InnerVolumeSpecName "kube-api-access-8ggwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:56:11 crc kubenswrapper[4744]: I1003 16:56:11.305892 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af14e991-2a54-4228-aa49-a8ce99d8eab2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "af14e991-2a54-4228-aa49-a8ce99d8eab2" (UID: "af14e991-2a54-4228-aa49-a8ce99d8eab2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:56:11 crc kubenswrapper[4744]: I1003 16:56:11.315046 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af14e991-2a54-4228-aa49-a8ce99d8eab2-inventory" (OuterVolumeSpecName: "inventory") pod "af14e991-2a54-4228-aa49-a8ce99d8eab2" (UID: "af14e991-2a54-4228-aa49-a8ce99d8eab2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:56:11 crc kubenswrapper[4744]: I1003 16:56:11.360319 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/af14e991-2a54-4228-aa49-a8ce99d8eab2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:56:11 crc kubenswrapper[4744]: I1003 16:56:11.360383 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ggwp\" (UniqueName: \"kubernetes.io/projected/af14e991-2a54-4228-aa49-a8ce99d8eab2-kube-api-access-8ggwp\") on node \"crc\" DevicePath \"\"" Oct 03 16:56:11 crc kubenswrapper[4744]: I1003 16:56:11.360398 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af14e991-2a54-4228-aa49-a8ce99d8eab2-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 16:56:11 crc kubenswrapper[4744]: I1003 16:56:11.617814 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" event={"ID":"af14e991-2a54-4228-aa49-a8ce99d8eab2","Type":"ContainerDied","Data":"22a7650ab93b617444293b706c863fd854b16c3037e27f691d0318fd762257ea"} Oct 03 16:56:11 crc kubenswrapper[4744]: I1003 16:56:11.617878 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22a7650ab93b617444293b706c863fd854b16c3037e27f691d0318fd762257ea" Oct 03 16:56:11 crc kubenswrapper[4744]: I1003 16:56:11.617906 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.032617 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c"] Oct 03 16:56:18 crc kubenswrapper[4744]: E1003 16:56:18.033428 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af14e991-2a54-4228-aa49-a8ce99d8eab2" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.033443 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="af14e991-2a54-4228-aa49-a8ce99d8eab2" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.033674 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="af14e991-2a54-4228-aa49-a8ce99d8eab2" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.035878 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.039476 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.040273 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.048885 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.054507 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.066449 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c"] Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.210407 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jffr\" (UniqueName: \"kubernetes.io/projected/2ac77289-8098-4cdd-892f-a56916c886e6-kube-api-access-2jffr\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c\" (UID: \"2ac77289-8098-4cdd-892f-a56916c886e6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.210541 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ac77289-8098-4cdd-892f-a56916c886e6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c\" (UID: \"2ac77289-8098-4cdd-892f-a56916c886e6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.211127 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ac77289-8098-4cdd-892f-a56916c886e6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c\" (UID: \"2ac77289-8098-4cdd-892f-a56916c886e6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.313156 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ac77289-8098-4cdd-892f-a56916c886e6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c\" (UID: \"2ac77289-8098-4cdd-892f-a56916c886e6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.313249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jffr\" (UniqueName: \"kubernetes.io/projected/2ac77289-8098-4cdd-892f-a56916c886e6-kube-api-access-2jffr\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c\" (UID: \"2ac77289-8098-4cdd-892f-a56916c886e6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.313287 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ac77289-8098-4cdd-892f-a56916c886e6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c\" (UID: \"2ac77289-8098-4cdd-892f-a56916c886e6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.321974 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ac77289-8098-4cdd-892f-a56916c886e6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c\" (UID: \"2ac77289-8098-4cdd-892f-a56916c886e6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.322686 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ac77289-8098-4cdd-892f-a56916c886e6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c\" (UID: \"2ac77289-8098-4cdd-892f-a56916c886e6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.336209 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jffr\" (UniqueName: \"kubernetes.io/projected/2ac77289-8098-4cdd-892f-a56916c886e6-kube-api-access-2jffr\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c\" (UID: \"2ac77289-8098-4cdd-892f-a56916c886e6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.381476 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" Oct 03 16:56:18 crc kubenswrapper[4744]: I1003 16:56:18.974967 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c"] Oct 03 16:56:19 crc kubenswrapper[4744]: I1003 16:56:19.716230 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" event={"ID":"2ac77289-8098-4cdd-892f-a56916c886e6","Type":"ContainerStarted","Data":"f2126584ddf7c92709e5d85c7a9437f1cef388d5af62a96adadf3b3fdc95f882"} Oct 03 16:56:20 crc kubenswrapper[4744]: I1003 16:56:20.733077 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" event={"ID":"2ac77289-8098-4cdd-892f-a56916c886e6","Type":"ContainerStarted","Data":"64dad4e1218f10ec3c8f2d6dd37acf1df431d8959420be872e4377d6d007ce08"} Oct 03 16:56:20 crc kubenswrapper[4744]: I1003 16:56:20.759776 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" podStartSLOduration=1.980256829 podStartE2EDuration="2.759760253s" podCreationTimestamp="2025-10-03 16:56:18 +0000 UTC" firstStartedPulling="2025-10-03 16:56:18.982888045 +0000 UTC m=+1905.262763991" lastFinishedPulling="2025-10-03 16:56:19.762391509 +0000 UTC m=+1906.042267415" observedRunningTime="2025-10-03 16:56:20.749335809 +0000 UTC m=+1907.029211745" watchObservedRunningTime="2025-10-03 16:56:20.759760253 +0000 UTC m=+1907.039636149" Oct 03 16:56:23 crc kubenswrapper[4744]: I1003 16:56:23.893262 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:56:23 crc kubenswrapper[4744]: E1003 16:56:23.894779 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 16:56:36 crc kubenswrapper[4744]: I1003 16:56:36.892678 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 16:56:37 crc kubenswrapper[4744]: I1003 16:56:37.953470 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"ffa178bfa89fc417d111dc4ef6ca5fa591b9352e2e2712e9cd221a0367de2cf0"} Oct 03 16:56:39 crc kubenswrapper[4744]: I1003 16:56:39.302393 4744 scope.go:117] "RemoveContainer" containerID="f51e3986742213bb93feed75ec3949a025047f40d33b489bdc97accde4460cab" Oct 03 16:56:39 crc kubenswrapper[4744]: I1003 16:56:39.378143 4744 scope.go:117] "RemoveContainer" containerID="18c2a82b8e79a2029579d897d7f01f0583fdfbd27b3aa482d08a4d429783f1a8" Oct 03 16:56:39 crc kubenswrapper[4744]: I1003 16:56:39.445066 4744 scope.go:117] "RemoveContainer" containerID="20a31bbd941873591b0299af896e8dd8cc01fae49345187d13a006528fb702a5" Oct 03 16:56:51 crc kubenswrapper[4744]: I1003 16:56:51.059530 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-h7gcm"] Oct 03 16:56:51 crc kubenswrapper[4744]: I1003 16:56:51.067602 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-h7gcm"] Oct 03 16:56:52 crc kubenswrapper[4744]: I1003 16:56:52.932132 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2cdd571-e0fc-40c3-bf1f-40bfcf91913d" path="/var/lib/kubelet/pods/e2cdd571-e0fc-40c3-bf1f-40bfcf91913d/volumes" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.276909 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9fxr5"] Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.280137 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.291831 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9fxr5"] Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.394732 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1726bb2c-cca1-4d32-b651-f3be55e815ef-catalog-content\") pod \"certified-operators-9fxr5\" (UID: \"1726bb2c-cca1-4d32-b651-f3be55e815ef\") " pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.394891 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1726bb2c-cca1-4d32-b651-f3be55e815ef-utilities\") pod \"certified-operators-9fxr5\" (UID: \"1726bb2c-cca1-4d32-b651-f3be55e815ef\") " pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.394916 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28gks\" (UniqueName: \"kubernetes.io/projected/1726bb2c-cca1-4d32-b651-f3be55e815ef-kube-api-access-28gks\") pod \"certified-operators-9fxr5\" (UID: \"1726bb2c-cca1-4d32-b651-f3be55e815ef\") " pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.472546 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4qj4r"] Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.474372 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.486839 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4qj4r"] Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.496455 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1726bb2c-cca1-4d32-b651-f3be55e815ef-utilities\") pod \"certified-operators-9fxr5\" (UID: \"1726bb2c-cca1-4d32-b651-f3be55e815ef\") " pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.496599 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28gks\" (UniqueName: \"kubernetes.io/projected/1726bb2c-cca1-4d32-b651-f3be55e815ef-kube-api-access-28gks\") pod \"certified-operators-9fxr5\" (UID: \"1726bb2c-cca1-4d32-b651-f3be55e815ef\") " pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.496698 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1726bb2c-cca1-4d32-b651-f3be55e815ef-catalog-content\") pod \"certified-operators-9fxr5\" (UID: \"1726bb2c-cca1-4d32-b651-f3be55e815ef\") " pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.497305 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1726bb2c-cca1-4d32-b651-f3be55e815ef-catalog-content\") pod \"certified-operators-9fxr5\" (UID: \"1726bb2c-cca1-4d32-b651-f3be55e815ef\") " pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.497659 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1726bb2c-cca1-4d32-b651-f3be55e815ef-utilities\") pod \"certified-operators-9fxr5\" (UID: \"1726bb2c-cca1-4d32-b651-f3be55e815ef\") " pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.525667 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28gks\" (UniqueName: \"kubernetes.io/projected/1726bb2c-cca1-4d32-b651-f3be55e815ef-kube-api-access-28gks\") pod \"certified-operators-9fxr5\" (UID: \"1726bb2c-cca1-4d32-b651-f3be55e815ef\") " pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.598251 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-utilities\") pod \"community-operators-4qj4r\" (UID: \"4e05d173-68d3-4ee9-81a7-1369b17e9cc4\") " pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.598757 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntv2d\" (UniqueName: \"kubernetes.io/projected/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-kube-api-access-ntv2d\") pod \"community-operators-4qj4r\" (UID: \"4e05d173-68d3-4ee9-81a7-1369b17e9cc4\") " pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.598928 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-catalog-content\") pod \"community-operators-4qj4r\" (UID: \"4e05d173-68d3-4ee9-81a7-1369b17e9cc4\") " pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.621871 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.701162 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-utilities\") pod \"community-operators-4qj4r\" (UID: \"4e05d173-68d3-4ee9-81a7-1369b17e9cc4\") " pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.701307 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntv2d\" (UniqueName: \"kubernetes.io/projected/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-kube-api-access-ntv2d\") pod \"community-operators-4qj4r\" (UID: \"4e05d173-68d3-4ee9-81a7-1369b17e9cc4\") " pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.701375 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-catalog-content\") pod \"community-operators-4qj4r\" (UID: \"4e05d173-68d3-4ee9-81a7-1369b17e9cc4\") " pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.701982 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-catalog-content\") pod \"community-operators-4qj4r\" (UID: \"4e05d173-68d3-4ee9-81a7-1369b17e9cc4\") " pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.702268 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-utilities\") pod \"community-operators-4qj4r\" (UID: \"4e05d173-68d3-4ee9-81a7-1369b17e9cc4\") " pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.735257 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntv2d\" (UniqueName: \"kubernetes.io/projected/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-kube-api-access-ntv2d\") pod \"community-operators-4qj4r\" (UID: \"4e05d173-68d3-4ee9-81a7-1369b17e9cc4\") " pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:56:57 crc kubenswrapper[4744]: I1003 16:56:57.799997 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:56:58 crc kubenswrapper[4744]: I1003 16:56:58.106826 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9fxr5"] Oct 03 16:56:58 crc kubenswrapper[4744]: W1003 16:56:58.117948 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1726bb2c_cca1_4d32_b651_f3be55e815ef.slice/crio-cc072b7a6e7b2270f07b7c9c606c50f4268eeaff035427a2779f12296661e0b4 WatchSource:0}: Error finding container cc072b7a6e7b2270f07b7c9c606c50f4268eeaff035427a2779f12296661e0b4: Status 404 returned error can't find the container with id cc072b7a6e7b2270f07b7c9c606c50f4268eeaff035427a2779f12296661e0b4 Oct 03 16:56:58 crc kubenswrapper[4744]: I1003 16:56:58.161930 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fxr5" event={"ID":"1726bb2c-cca1-4d32-b651-f3be55e815ef","Type":"ContainerStarted","Data":"cc072b7a6e7b2270f07b7c9c606c50f4268eeaff035427a2779f12296661e0b4"} Oct 03 16:56:58 crc kubenswrapper[4744]: I1003 16:56:58.325682 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4qj4r"] Oct 03 16:56:59 crc kubenswrapper[4744]: I1003 16:56:59.174860 4744 generic.go:334] "Generic (PLEG): container finished" podID="4e05d173-68d3-4ee9-81a7-1369b17e9cc4" containerID="4594bb57faed9af380aef0b4750d8f2b3fddd4f7acf2963a0ee6c354d1fd93ad" exitCode=0 Oct 03 16:56:59 crc kubenswrapper[4744]: I1003 16:56:59.175083 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qj4r" event={"ID":"4e05d173-68d3-4ee9-81a7-1369b17e9cc4","Type":"ContainerDied","Data":"4594bb57faed9af380aef0b4750d8f2b3fddd4f7acf2963a0ee6c354d1fd93ad"} Oct 03 16:56:59 crc kubenswrapper[4744]: I1003 16:56:59.175251 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qj4r" event={"ID":"4e05d173-68d3-4ee9-81a7-1369b17e9cc4","Type":"ContainerStarted","Data":"cdee60055ce997aebfc52d00bc106cae21ea77082e257afc502d18d640be3629"} Oct 03 16:56:59 crc kubenswrapper[4744]: I1003 16:56:59.177206 4744 generic.go:334] "Generic (PLEG): container finished" podID="1726bb2c-cca1-4d32-b651-f3be55e815ef" containerID="0996d467275a00f01024cb544426b096f362672023be2c6a43cf8bfa18624e02" exitCode=0 Oct 03 16:56:59 crc kubenswrapper[4744]: I1003 16:56:59.177238 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fxr5" event={"ID":"1726bb2c-cca1-4d32-b651-f3be55e815ef","Type":"ContainerDied","Data":"0996d467275a00f01024cb544426b096f362672023be2c6a43cf8bfa18624e02"} Oct 03 16:56:59 crc kubenswrapper[4744]: I1003 16:56:59.179406 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 16:57:01 crc kubenswrapper[4744]: I1003 16:57:01.193889 4744 generic.go:334] "Generic (PLEG): container finished" podID="4e05d173-68d3-4ee9-81a7-1369b17e9cc4" containerID="b1e143ab8da7146c6292d7bc1a63bcd8256480faf79ea8c6a26023749be34c19" exitCode=0 Oct 03 16:57:01 crc kubenswrapper[4744]: I1003 16:57:01.193931 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qj4r" event={"ID":"4e05d173-68d3-4ee9-81a7-1369b17e9cc4","Type":"ContainerDied","Data":"b1e143ab8da7146c6292d7bc1a63bcd8256480faf79ea8c6a26023749be34c19"} Oct 03 16:57:01 crc kubenswrapper[4744]: I1003 16:57:01.196921 4744 generic.go:334] "Generic (PLEG): container finished" podID="1726bb2c-cca1-4d32-b651-f3be55e815ef" containerID="bfd9bc21498b5f99d8806c22b9ba186d6bb0f44c0ee6b886ed20d492d8087a1d" exitCode=0 Oct 03 16:57:01 crc kubenswrapper[4744]: I1003 16:57:01.196966 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fxr5" event={"ID":"1726bb2c-cca1-4d32-b651-f3be55e815ef","Type":"ContainerDied","Data":"bfd9bc21498b5f99d8806c22b9ba186d6bb0f44c0ee6b886ed20d492d8087a1d"} Oct 03 16:57:02 crc kubenswrapper[4744]: I1003 16:57:02.210698 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fxr5" event={"ID":"1726bb2c-cca1-4d32-b651-f3be55e815ef","Type":"ContainerStarted","Data":"e5a8d3632060290abbfcdefa11cac25417c8a4d9cae4b29020851c376c7aace9"} Oct 03 16:57:02 crc kubenswrapper[4744]: I1003 16:57:02.236374 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9fxr5" podStartSLOduration=2.754284948 podStartE2EDuration="5.236357097s" podCreationTimestamp="2025-10-03 16:56:57 +0000 UTC" firstStartedPulling="2025-10-03 16:56:59.180130196 +0000 UTC m=+1945.460006082" lastFinishedPulling="2025-10-03 16:57:01.662202335 +0000 UTC m=+1947.942078231" observedRunningTime="2025-10-03 16:57:02.228100388 +0000 UTC m=+1948.507976294" watchObservedRunningTime="2025-10-03 16:57:02.236357097 +0000 UTC m=+1948.516232993" Oct 03 16:57:03 crc kubenswrapper[4744]: I1003 16:57:03.221428 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qj4r" event={"ID":"4e05d173-68d3-4ee9-81a7-1369b17e9cc4","Type":"ContainerStarted","Data":"daaaa10e78fd51d14e7b94618e3b1a6fc54769c7c7d54328f0ce4dfc2070d31c"} Oct 03 16:57:03 crc kubenswrapper[4744]: I1003 16:57:03.246485 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4qj4r" podStartSLOduration=3.347981792 podStartE2EDuration="6.246464131s" podCreationTimestamp="2025-10-03 16:56:57 +0000 UTC" firstStartedPulling="2025-10-03 16:56:59.179028589 +0000 UTC m=+1945.458904495" lastFinishedPulling="2025-10-03 16:57:02.077510938 +0000 UTC m=+1948.357386834" observedRunningTime="2025-10-03 16:57:03.244092202 +0000 UTC m=+1949.523968128" watchObservedRunningTime="2025-10-03 16:57:03.246464131 +0000 UTC m=+1949.526340027" Oct 03 16:57:04 crc kubenswrapper[4744]: E1003 16:57:04.560242 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ac77289_8098_4cdd_892f_a56916c886e6.slice/crio-conmon-64dad4e1218f10ec3c8f2d6dd37acf1df431d8959420be872e4377d6d007ce08.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ac77289_8098_4cdd_892f_a56916c886e6.slice/crio-64dad4e1218f10ec3c8f2d6dd37acf1df431d8959420be872e4377d6d007ce08.scope\": RecentStats: unable to find data in memory cache]" Oct 03 16:57:05 crc kubenswrapper[4744]: I1003 16:57:05.242420 4744 generic.go:334] "Generic (PLEG): container finished" podID="2ac77289-8098-4cdd-892f-a56916c886e6" containerID="64dad4e1218f10ec3c8f2d6dd37acf1df431d8959420be872e4377d6d007ce08" exitCode=0 Oct 03 16:57:05 crc kubenswrapper[4744]: I1003 16:57:05.242480 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" event={"ID":"2ac77289-8098-4cdd-892f-a56916c886e6","Type":"ContainerDied","Data":"64dad4e1218f10ec3c8f2d6dd37acf1df431d8959420be872e4377d6d007ce08"} Oct 03 16:57:06 crc kubenswrapper[4744]: I1003 16:57:06.661191 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" Oct 03 16:57:06 crc kubenswrapper[4744]: I1003 16:57:06.806559 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ac77289-8098-4cdd-892f-a56916c886e6-inventory\") pod \"2ac77289-8098-4cdd-892f-a56916c886e6\" (UID: \"2ac77289-8098-4cdd-892f-a56916c886e6\") " Oct 03 16:57:06 crc kubenswrapper[4744]: I1003 16:57:06.806676 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jffr\" (UniqueName: \"kubernetes.io/projected/2ac77289-8098-4cdd-892f-a56916c886e6-kube-api-access-2jffr\") pod \"2ac77289-8098-4cdd-892f-a56916c886e6\" (UID: \"2ac77289-8098-4cdd-892f-a56916c886e6\") " Oct 03 16:57:06 crc kubenswrapper[4744]: I1003 16:57:06.806759 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ac77289-8098-4cdd-892f-a56916c886e6-ssh-key\") pod \"2ac77289-8098-4cdd-892f-a56916c886e6\" (UID: \"2ac77289-8098-4cdd-892f-a56916c886e6\") " Oct 03 16:57:06 crc kubenswrapper[4744]: I1003 16:57:06.814763 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ac77289-8098-4cdd-892f-a56916c886e6-kube-api-access-2jffr" (OuterVolumeSpecName: "kube-api-access-2jffr") pod "2ac77289-8098-4cdd-892f-a56916c886e6" (UID: "2ac77289-8098-4cdd-892f-a56916c886e6"). InnerVolumeSpecName "kube-api-access-2jffr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:57:06 crc kubenswrapper[4744]: I1003 16:57:06.845842 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ac77289-8098-4cdd-892f-a56916c886e6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2ac77289-8098-4cdd-892f-a56916c886e6" (UID: "2ac77289-8098-4cdd-892f-a56916c886e6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:57:06 crc kubenswrapper[4744]: I1003 16:57:06.850167 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ac77289-8098-4cdd-892f-a56916c886e6-inventory" (OuterVolumeSpecName: "inventory") pod "2ac77289-8098-4cdd-892f-a56916c886e6" (UID: "2ac77289-8098-4cdd-892f-a56916c886e6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:57:06 crc kubenswrapper[4744]: I1003 16:57:06.909679 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ac77289-8098-4cdd-892f-a56916c886e6-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:06 crc kubenswrapper[4744]: I1003 16:57:06.909881 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jffr\" (UniqueName: \"kubernetes.io/projected/2ac77289-8098-4cdd-892f-a56916c886e6-kube-api-access-2jffr\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:06 crc kubenswrapper[4744]: I1003 16:57:06.909957 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ac77289-8098-4cdd-892f-a56916c886e6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.264954 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" event={"ID":"2ac77289-8098-4cdd-892f-a56916c886e6","Type":"ContainerDied","Data":"f2126584ddf7c92709e5d85c7a9437f1cef388d5af62a96adadf3b3fdc95f882"} Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.265009 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2126584ddf7c92709e5d85c7a9437f1cef388d5af62a96adadf3b3fdc95f882" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.265021 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.366939 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8nbg9"] Oct 03 16:57:07 crc kubenswrapper[4744]: E1003 16:57:07.367416 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ac77289-8098-4cdd-892f-a56916c886e6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.367440 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ac77289-8098-4cdd-892f-a56916c886e6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.367726 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ac77289-8098-4cdd-892f-a56916c886e6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.368528 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.370972 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.371309 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.371847 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.372032 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.381347 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8nbg9"] Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.420819 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1700fe09-1277-4cda-89dc-03d220941b9a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8nbg9\" (UID: \"1700fe09-1277-4cda-89dc-03d220941b9a\") " pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.420878 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1700fe09-1277-4cda-89dc-03d220941b9a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8nbg9\" (UID: \"1700fe09-1277-4cda-89dc-03d220941b9a\") " pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.421298 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s44pd\" (UniqueName: \"kubernetes.io/projected/1700fe09-1277-4cda-89dc-03d220941b9a-kube-api-access-s44pd\") pod \"ssh-known-hosts-edpm-deployment-8nbg9\" (UID: \"1700fe09-1277-4cda-89dc-03d220941b9a\") " pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.522954 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1700fe09-1277-4cda-89dc-03d220941b9a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8nbg9\" (UID: \"1700fe09-1277-4cda-89dc-03d220941b9a\") " pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.523030 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1700fe09-1277-4cda-89dc-03d220941b9a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8nbg9\" (UID: \"1700fe09-1277-4cda-89dc-03d220941b9a\") " pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.523305 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s44pd\" (UniqueName: \"kubernetes.io/projected/1700fe09-1277-4cda-89dc-03d220941b9a-kube-api-access-s44pd\") pod \"ssh-known-hosts-edpm-deployment-8nbg9\" (UID: \"1700fe09-1277-4cda-89dc-03d220941b9a\") " pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.528750 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1700fe09-1277-4cda-89dc-03d220941b9a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8nbg9\" (UID: \"1700fe09-1277-4cda-89dc-03d220941b9a\") " pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.530540 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1700fe09-1277-4cda-89dc-03d220941b9a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8nbg9\" (UID: \"1700fe09-1277-4cda-89dc-03d220941b9a\") " pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.540888 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s44pd\" (UniqueName: \"kubernetes.io/projected/1700fe09-1277-4cda-89dc-03d220941b9a-kube-api-access-s44pd\") pod \"ssh-known-hosts-edpm-deployment-8nbg9\" (UID: \"1700fe09-1277-4cda-89dc-03d220941b9a\") " pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.623021 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.624743 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.678640 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.689801 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.800226 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.800579 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:57:07 crc kubenswrapper[4744]: I1003 16:57:07.932765 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:57:08 crc kubenswrapper[4744]: I1003 16:57:08.215298 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8nbg9"] Oct 03 16:57:08 crc kubenswrapper[4744]: I1003 16:57:08.272650 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" event={"ID":"1700fe09-1277-4cda-89dc-03d220941b9a","Type":"ContainerStarted","Data":"dc546dd5137f9f4e3c0b1002fa4632dae379b5c2f06eecb92543231920fd3d79"} Oct 03 16:57:08 crc kubenswrapper[4744]: I1003 16:57:08.325088 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:57:08 crc kubenswrapper[4744]: I1003 16:57:08.325818 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:57:09 crc kubenswrapper[4744]: I1003 16:57:09.287391 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" event={"ID":"1700fe09-1277-4cda-89dc-03d220941b9a","Type":"ContainerStarted","Data":"01de597ed590961b0c07e9d59ed32eaf6822deaabfc29da84674e62d25d0a36c"} Oct 03 16:57:09 crc kubenswrapper[4744]: I1003 16:57:09.313641 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" podStartSLOduration=1.746567846 podStartE2EDuration="2.313612658s" podCreationTimestamp="2025-10-03 16:57:07 +0000 UTC" firstStartedPulling="2025-10-03 16:57:08.220316713 +0000 UTC m=+1954.500192609" lastFinishedPulling="2025-10-03 16:57:08.787361515 +0000 UTC m=+1955.067237421" observedRunningTime="2025-10-03 16:57:09.308369405 +0000 UTC m=+1955.588245341" watchObservedRunningTime="2025-10-03 16:57:09.313612658 +0000 UTC m=+1955.593488594" Oct 03 16:57:09 crc kubenswrapper[4744]: I1003 16:57:09.660456 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9fxr5"] Oct 03 16:57:10 crc kubenswrapper[4744]: I1003 16:57:10.295397 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9fxr5" podUID="1726bb2c-cca1-4d32-b651-f3be55e815ef" containerName="registry-server" containerID="cri-o://e5a8d3632060290abbfcdefa11cac25417c8a4d9cae4b29020851c376c7aace9" gracePeriod=2 Oct 03 16:57:10 crc kubenswrapper[4744]: I1003 16:57:10.658765 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4qj4r"] Oct 03 16:57:10 crc kubenswrapper[4744]: I1003 16:57:10.832214 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:57:10 crc kubenswrapper[4744]: I1003 16:57:10.902895 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1726bb2c-cca1-4d32-b651-f3be55e815ef-utilities\") pod \"1726bb2c-cca1-4d32-b651-f3be55e815ef\" (UID: \"1726bb2c-cca1-4d32-b651-f3be55e815ef\") " Oct 03 16:57:10 crc kubenswrapper[4744]: I1003 16:57:10.903126 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28gks\" (UniqueName: \"kubernetes.io/projected/1726bb2c-cca1-4d32-b651-f3be55e815ef-kube-api-access-28gks\") pod \"1726bb2c-cca1-4d32-b651-f3be55e815ef\" (UID: \"1726bb2c-cca1-4d32-b651-f3be55e815ef\") " Oct 03 16:57:10 crc kubenswrapper[4744]: I1003 16:57:10.903195 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1726bb2c-cca1-4d32-b651-f3be55e815ef-catalog-content\") pod \"1726bb2c-cca1-4d32-b651-f3be55e815ef\" (UID: \"1726bb2c-cca1-4d32-b651-f3be55e815ef\") " Oct 03 16:57:10 crc kubenswrapper[4744]: I1003 16:57:10.905186 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1726bb2c-cca1-4d32-b651-f3be55e815ef-utilities" (OuterVolumeSpecName: "utilities") pod "1726bb2c-cca1-4d32-b651-f3be55e815ef" (UID: "1726bb2c-cca1-4d32-b651-f3be55e815ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:57:10 crc kubenswrapper[4744]: I1003 16:57:10.910775 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1726bb2c-cca1-4d32-b651-f3be55e815ef-kube-api-access-28gks" (OuterVolumeSpecName: "kube-api-access-28gks") pod "1726bb2c-cca1-4d32-b651-f3be55e815ef" (UID: "1726bb2c-cca1-4d32-b651-f3be55e815ef"). InnerVolumeSpecName "kube-api-access-28gks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:57:10 crc kubenswrapper[4744]: I1003 16:57:10.970685 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1726bb2c-cca1-4d32-b651-f3be55e815ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1726bb2c-cca1-4d32-b651-f3be55e815ef" (UID: "1726bb2c-cca1-4d32-b651-f3be55e815ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.005972 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1726bb2c-cca1-4d32-b651-f3be55e815ef-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.006016 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1726bb2c-cca1-4d32-b651-f3be55e815ef-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.006029 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28gks\" (UniqueName: \"kubernetes.io/projected/1726bb2c-cca1-4d32-b651-f3be55e815ef-kube-api-access-28gks\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.306555 4744 generic.go:334] "Generic (PLEG): container finished" podID="1726bb2c-cca1-4d32-b651-f3be55e815ef" containerID="e5a8d3632060290abbfcdefa11cac25417c8a4d9cae4b29020851c376c7aace9" exitCode=0 Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.306610 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9fxr5" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.306624 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fxr5" event={"ID":"1726bb2c-cca1-4d32-b651-f3be55e815ef","Type":"ContainerDied","Data":"e5a8d3632060290abbfcdefa11cac25417c8a4d9cae4b29020851c376c7aace9"} Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.308122 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9fxr5" event={"ID":"1726bb2c-cca1-4d32-b651-f3be55e815ef","Type":"ContainerDied","Data":"cc072b7a6e7b2270f07b7c9c606c50f4268eeaff035427a2779f12296661e0b4"} Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.308152 4744 scope.go:117] "RemoveContainer" containerID="e5a8d3632060290abbfcdefa11cac25417c8a4d9cae4b29020851c376c7aace9" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.308696 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4qj4r" podUID="4e05d173-68d3-4ee9-81a7-1369b17e9cc4" containerName="registry-server" containerID="cri-o://daaaa10e78fd51d14e7b94618e3b1a6fc54769c7c7d54328f0ce4dfc2070d31c" gracePeriod=2 Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.333444 4744 scope.go:117] "RemoveContainer" containerID="bfd9bc21498b5f99d8806c22b9ba186d6bb0f44c0ee6b886ed20d492d8087a1d" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.353869 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9fxr5"] Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.370529 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9fxr5"] Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.389694 4744 scope.go:117] "RemoveContainer" containerID="0996d467275a00f01024cb544426b096f362672023be2c6a43cf8bfa18624e02" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.571054 4744 scope.go:117] "RemoveContainer" containerID="e5a8d3632060290abbfcdefa11cac25417c8a4d9cae4b29020851c376c7aace9" Oct 03 16:57:11 crc kubenswrapper[4744]: E1003 16:57:11.571848 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5a8d3632060290abbfcdefa11cac25417c8a4d9cae4b29020851c376c7aace9\": container with ID starting with e5a8d3632060290abbfcdefa11cac25417c8a4d9cae4b29020851c376c7aace9 not found: ID does not exist" containerID="e5a8d3632060290abbfcdefa11cac25417c8a4d9cae4b29020851c376c7aace9" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.571896 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5a8d3632060290abbfcdefa11cac25417c8a4d9cae4b29020851c376c7aace9"} err="failed to get container status \"e5a8d3632060290abbfcdefa11cac25417c8a4d9cae4b29020851c376c7aace9\": rpc error: code = NotFound desc = could not find container \"e5a8d3632060290abbfcdefa11cac25417c8a4d9cae4b29020851c376c7aace9\": container with ID starting with e5a8d3632060290abbfcdefa11cac25417c8a4d9cae4b29020851c376c7aace9 not found: ID does not exist" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.571930 4744 scope.go:117] "RemoveContainer" containerID="bfd9bc21498b5f99d8806c22b9ba186d6bb0f44c0ee6b886ed20d492d8087a1d" Oct 03 16:57:11 crc kubenswrapper[4744]: E1003 16:57:11.572437 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfd9bc21498b5f99d8806c22b9ba186d6bb0f44c0ee6b886ed20d492d8087a1d\": container with ID starting with bfd9bc21498b5f99d8806c22b9ba186d6bb0f44c0ee6b886ed20d492d8087a1d not found: ID does not exist" containerID="bfd9bc21498b5f99d8806c22b9ba186d6bb0f44c0ee6b886ed20d492d8087a1d" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.572481 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfd9bc21498b5f99d8806c22b9ba186d6bb0f44c0ee6b886ed20d492d8087a1d"} err="failed to get container status \"bfd9bc21498b5f99d8806c22b9ba186d6bb0f44c0ee6b886ed20d492d8087a1d\": rpc error: code = NotFound desc = could not find container \"bfd9bc21498b5f99d8806c22b9ba186d6bb0f44c0ee6b886ed20d492d8087a1d\": container with ID starting with bfd9bc21498b5f99d8806c22b9ba186d6bb0f44c0ee6b886ed20d492d8087a1d not found: ID does not exist" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.572518 4744 scope.go:117] "RemoveContainer" containerID="0996d467275a00f01024cb544426b096f362672023be2c6a43cf8bfa18624e02" Oct 03 16:57:11 crc kubenswrapper[4744]: E1003 16:57:11.572866 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0996d467275a00f01024cb544426b096f362672023be2c6a43cf8bfa18624e02\": container with ID starting with 0996d467275a00f01024cb544426b096f362672023be2c6a43cf8bfa18624e02 not found: ID does not exist" containerID="0996d467275a00f01024cb544426b096f362672023be2c6a43cf8bfa18624e02" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.572910 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0996d467275a00f01024cb544426b096f362672023be2c6a43cf8bfa18624e02"} err="failed to get container status \"0996d467275a00f01024cb544426b096f362672023be2c6a43cf8bfa18624e02\": rpc error: code = NotFound desc = could not find container \"0996d467275a00f01024cb544426b096f362672023be2c6a43cf8bfa18624e02\": container with ID starting with 0996d467275a00f01024cb544426b096f362672023be2c6a43cf8bfa18624e02 not found: ID does not exist" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.816553 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.929171 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-utilities\") pod \"4e05d173-68d3-4ee9-81a7-1369b17e9cc4\" (UID: \"4e05d173-68d3-4ee9-81a7-1369b17e9cc4\") " Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.929314 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntv2d\" (UniqueName: \"kubernetes.io/projected/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-kube-api-access-ntv2d\") pod \"4e05d173-68d3-4ee9-81a7-1369b17e9cc4\" (UID: \"4e05d173-68d3-4ee9-81a7-1369b17e9cc4\") " Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.929556 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-catalog-content\") pod \"4e05d173-68d3-4ee9-81a7-1369b17e9cc4\" (UID: \"4e05d173-68d3-4ee9-81a7-1369b17e9cc4\") " Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.932651 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-utilities" (OuterVolumeSpecName: "utilities") pod "4e05d173-68d3-4ee9-81a7-1369b17e9cc4" (UID: "4e05d173-68d3-4ee9-81a7-1369b17e9cc4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:57:11 crc kubenswrapper[4744]: I1003 16:57:11.939504 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-kube-api-access-ntv2d" (OuterVolumeSpecName: "kube-api-access-ntv2d") pod "4e05d173-68d3-4ee9-81a7-1369b17e9cc4" (UID: "4e05d173-68d3-4ee9-81a7-1369b17e9cc4"). InnerVolumeSpecName "kube-api-access-ntv2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.000509 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e05d173-68d3-4ee9-81a7-1369b17e9cc4" (UID: "4e05d173-68d3-4ee9-81a7-1369b17e9cc4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.032682 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.032717 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntv2d\" (UniqueName: \"kubernetes.io/projected/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-kube-api-access-ntv2d\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.032728 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e05d173-68d3-4ee9-81a7-1369b17e9cc4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.322315 4744 generic.go:334] "Generic (PLEG): container finished" podID="4e05d173-68d3-4ee9-81a7-1369b17e9cc4" containerID="daaaa10e78fd51d14e7b94618e3b1a6fc54769c7c7d54328f0ce4dfc2070d31c" exitCode=0 Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.322380 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qj4r" event={"ID":"4e05d173-68d3-4ee9-81a7-1369b17e9cc4","Type":"ContainerDied","Data":"daaaa10e78fd51d14e7b94618e3b1a6fc54769c7c7d54328f0ce4dfc2070d31c"} Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.322406 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4qj4r" event={"ID":"4e05d173-68d3-4ee9-81a7-1369b17e9cc4","Type":"ContainerDied","Data":"cdee60055ce997aebfc52d00bc106cae21ea77082e257afc502d18d640be3629"} Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.322422 4744 scope.go:117] "RemoveContainer" containerID="daaaa10e78fd51d14e7b94618e3b1a6fc54769c7c7d54328f0ce4dfc2070d31c" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.323607 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4qj4r" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.353354 4744 scope.go:117] "RemoveContainer" containerID="b1e143ab8da7146c6292d7bc1a63bcd8256480faf79ea8c6a26023749be34c19" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.385555 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4qj4r"] Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.400139 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4qj4r"] Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.410876 4744 scope.go:117] "RemoveContainer" containerID="4594bb57faed9af380aef0b4750d8f2b3fddd4f7acf2963a0ee6c354d1fd93ad" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.430035 4744 scope.go:117] "RemoveContainer" containerID="daaaa10e78fd51d14e7b94618e3b1a6fc54769c7c7d54328f0ce4dfc2070d31c" Oct 03 16:57:12 crc kubenswrapper[4744]: E1003 16:57:12.436021 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daaaa10e78fd51d14e7b94618e3b1a6fc54769c7c7d54328f0ce4dfc2070d31c\": container with ID starting with daaaa10e78fd51d14e7b94618e3b1a6fc54769c7c7d54328f0ce4dfc2070d31c not found: ID does not exist" containerID="daaaa10e78fd51d14e7b94618e3b1a6fc54769c7c7d54328f0ce4dfc2070d31c" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.436083 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daaaa10e78fd51d14e7b94618e3b1a6fc54769c7c7d54328f0ce4dfc2070d31c"} err="failed to get container status \"daaaa10e78fd51d14e7b94618e3b1a6fc54769c7c7d54328f0ce4dfc2070d31c\": rpc error: code = NotFound desc = could not find container \"daaaa10e78fd51d14e7b94618e3b1a6fc54769c7c7d54328f0ce4dfc2070d31c\": container with ID starting with daaaa10e78fd51d14e7b94618e3b1a6fc54769c7c7d54328f0ce4dfc2070d31c not found: ID does not exist" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.436107 4744 scope.go:117] "RemoveContainer" containerID="b1e143ab8da7146c6292d7bc1a63bcd8256480faf79ea8c6a26023749be34c19" Oct 03 16:57:12 crc kubenswrapper[4744]: E1003 16:57:12.436510 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1e143ab8da7146c6292d7bc1a63bcd8256480faf79ea8c6a26023749be34c19\": container with ID starting with b1e143ab8da7146c6292d7bc1a63bcd8256480faf79ea8c6a26023749be34c19 not found: ID does not exist" containerID="b1e143ab8da7146c6292d7bc1a63bcd8256480faf79ea8c6a26023749be34c19" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.436552 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1e143ab8da7146c6292d7bc1a63bcd8256480faf79ea8c6a26023749be34c19"} err="failed to get container status \"b1e143ab8da7146c6292d7bc1a63bcd8256480faf79ea8c6a26023749be34c19\": rpc error: code = NotFound desc = could not find container \"b1e143ab8da7146c6292d7bc1a63bcd8256480faf79ea8c6a26023749be34c19\": container with ID starting with b1e143ab8da7146c6292d7bc1a63bcd8256480faf79ea8c6a26023749be34c19 not found: ID does not exist" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.436580 4744 scope.go:117] "RemoveContainer" containerID="4594bb57faed9af380aef0b4750d8f2b3fddd4f7acf2963a0ee6c354d1fd93ad" Oct 03 16:57:12 crc kubenswrapper[4744]: E1003 16:57:12.436952 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4594bb57faed9af380aef0b4750d8f2b3fddd4f7acf2963a0ee6c354d1fd93ad\": container with ID starting with 4594bb57faed9af380aef0b4750d8f2b3fddd4f7acf2963a0ee6c354d1fd93ad not found: ID does not exist" containerID="4594bb57faed9af380aef0b4750d8f2b3fddd4f7acf2963a0ee6c354d1fd93ad" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.436971 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4594bb57faed9af380aef0b4750d8f2b3fddd4f7acf2963a0ee6c354d1fd93ad"} err="failed to get container status \"4594bb57faed9af380aef0b4750d8f2b3fddd4f7acf2963a0ee6c354d1fd93ad\": rpc error: code = NotFound desc = could not find container \"4594bb57faed9af380aef0b4750d8f2b3fddd4f7acf2963a0ee6c354d1fd93ad\": container with ID starting with 4594bb57faed9af380aef0b4750d8f2b3fddd4f7acf2963a0ee6c354d1fd93ad not found: ID does not exist" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.909731 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1726bb2c-cca1-4d32-b651-f3be55e815ef" path="/var/lib/kubelet/pods/1726bb2c-cca1-4d32-b651-f3be55e815ef/volumes" Oct 03 16:57:12 crc kubenswrapper[4744]: I1003 16:57:12.912113 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e05d173-68d3-4ee9-81a7-1369b17e9cc4" path="/var/lib/kubelet/pods/4e05d173-68d3-4ee9-81a7-1369b17e9cc4/volumes" Oct 03 16:57:16 crc kubenswrapper[4744]: I1003 16:57:16.364393 4744 generic.go:334] "Generic (PLEG): container finished" podID="1700fe09-1277-4cda-89dc-03d220941b9a" containerID="01de597ed590961b0c07e9d59ed32eaf6822deaabfc29da84674e62d25d0a36c" exitCode=0 Oct 03 16:57:16 crc kubenswrapper[4744]: I1003 16:57:16.364410 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" event={"ID":"1700fe09-1277-4cda-89dc-03d220941b9a","Type":"ContainerDied","Data":"01de597ed590961b0c07e9d59ed32eaf6822deaabfc29da84674e62d25d0a36c"} Oct 03 16:57:17 crc kubenswrapper[4744]: I1003 16:57:17.886401 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" Oct 03 16:57:17 crc kubenswrapper[4744]: I1003 16:57:17.952991 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s44pd\" (UniqueName: \"kubernetes.io/projected/1700fe09-1277-4cda-89dc-03d220941b9a-kube-api-access-s44pd\") pod \"1700fe09-1277-4cda-89dc-03d220941b9a\" (UID: \"1700fe09-1277-4cda-89dc-03d220941b9a\") " Oct 03 16:57:17 crc kubenswrapper[4744]: I1003 16:57:17.953112 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1700fe09-1277-4cda-89dc-03d220941b9a-ssh-key-openstack-edpm-ipam\") pod \"1700fe09-1277-4cda-89dc-03d220941b9a\" (UID: \"1700fe09-1277-4cda-89dc-03d220941b9a\") " Oct 03 16:57:17 crc kubenswrapper[4744]: I1003 16:57:17.953173 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1700fe09-1277-4cda-89dc-03d220941b9a-inventory-0\") pod \"1700fe09-1277-4cda-89dc-03d220941b9a\" (UID: \"1700fe09-1277-4cda-89dc-03d220941b9a\") " Oct 03 16:57:17 crc kubenswrapper[4744]: I1003 16:57:17.960026 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1700fe09-1277-4cda-89dc-03d220941b9a-kube-api-access-s44pd" (OuterVolumeSpecName: "kube-api-access-s44pd") pod "1700fe09-1277-4cda-89dc-03d220941b9a" (UID: "1700fe09-1277-4cda-89dc-03d220941b9a"). InnerVolumeSpecName "kube-api-access-s44pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:57:17 crc kubenswrapper[4744]: I1003 16:57:17.989399 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1700fe09-1277-4cda-89dc-03d220941b9a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1700fe09-1277-4cda-89dc-03d220941b9a" (UID: "1700fe09-1277-4cda-89dc-03d220941b9a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:57:17 crc kubenswrapper[4744]: I1003 16:57:17.998594 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1700fe09-1277-4cda-89dc-03d220941b9a-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "1700fe09-1277-4cda-89dc-03d220941b9a" (UID: "1700fe09-1277-4cda-89dc-03d220941b9a"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.055554 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s44pd\" (UniqueName: \"kubernetes.io/projected/1700fe09-1277-4cda-89dc-03d220941b9a-kube-api-access-s44pd\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.055581 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1700fe09-1277-4cda-89dc-03d220941b9a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.055594 4744 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1700fe09-1277-4cda-89dc-03d220941b9a-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.396622 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" event={"ID":"1700fe09-1277-4cda-89dc-03d220941b9a","Type":"ContainerDied","Data":"dc546dd5137f9f4e3c0b1002fa4632dae379b5c2f06eecb92543231920fd3d79"} Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.397074 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc546dd5137f9f4e3c0b1002fa4632dae379b5c2f06eecb92543231920fd3d79" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.396847 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8nbg9" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.558677 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj"] Oct 03 16:57:18 crc kubenswrapper[4744]: E1003 16:57:18.559090 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e05d173-68d3-4ee9-81a7-1369b17e9cc4" containerName="extract-content" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.559111 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e05d173-68d3-4ee9-81a7-1369b17e9cc4" containerName="extract-content" Oct 03 16:57:18 crc kubenswrapper[4744]: E1003 16:57:18.559128 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e05d173-68d3-4ee9-81a7-1369b17e9cc4" containerName="extract-utilities" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.559138 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e05d173-68d3-4ee9-81a7-1369b17e9cc4" containerName="extract-utilities" Oct 03 16:57:18 crc kubenswrapper[4744]: E1003 16:57:18.559160 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1726bb2c-cca1-4d32-b651-f3be55e815ef" containerName="extract-utilities" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.559170 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1726bb2c-cca1-4d32-b651-f3be55e815ef" containerName="extract-utilities" Oct 03 16:57:18 crc kubenswrapper[4744]: E1003 16:57:18.559203 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1726bb2c-cca1-4d32-b651-f3be55e815ef" containerName="extract-content" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.559211 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1726bb2c-cca1-4d32-b651-f3be55e815ef" containerName="extract-content" Oct 03 16:57:18 crc kubenswrapper[4744]: E1003 16:57:18.559224 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1726bb2c-cca1-4d32-b651-f3be55e815ef" containerName="registry-server" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.559232 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1726bb2c-cca1-4d32-b651-f3be55e815ef" containerName="registry-server" Oct 03 16:57:18 crc kubenswrapper[4744]: E1003 16:57:18.559242 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e05d173-68d3-4ee9-81a7-1369b17e9cc4" containerName="registry-server" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.559249 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e05d173-68d3-4ee9-81a7-1369b17e9cc4" containerName="registry-server" Oct 03 16:57:18 crc kubenswrapper[4744]: E1003 16:57:18.559272 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1700fe09-1277-4cda-89dc-03d220941b9a" containerName="ssh-known-hosts-edpm-deployment" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.559280 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1700fe09-1277-4cda-89dc-03d220941b9a" containerName="ssh-known-hosts-edpm-deployment" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.559490 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1726bb2c-cca1-4d32-b651-f3be55e815ef" containerName="registry-server" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.559552 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1700fe09-1277-4cda-89dc-03d220941b9a" containerName="ssh-known-hosts-edpm-deployment" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.559570 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e05d173-68d3-4ee9-81a7-1369b17e9cc4" containerName="registry-server" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.560248 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.563091 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.564257 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.564732 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.565399 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.572447 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj"] Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.669661 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b94f190-a56d-420d-9558-7ec1116b90b9-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mf5dj\" (UID: \"1b94f190-a56d-420d-9558-7ec1116b90b9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.669724 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b94f190-a56d-420d-9558-7ec1116b90b9-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mf5dj\" (UID: \"1b94f190-a56d-420d-9558-7ec1116b90b9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.669955 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzz4t\" (UniqueName: \"kubernetes.io/projected/1b94f190-a56d-420d-9558-7ec1116b90b9-kube-api-access-vzz4t\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mf5dj\" (UID: \"1b94f190-a56d-420d-9558-7ec1116b90b9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.772002 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b94f190-a56d-420d-9558-7ec1116b90b9-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mf5dj\" (UID: \"1b94f190-a56d-420d-9558-7ec1116b90b9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.772058 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b94f190-a56d-420d-9558-7ec1116b90b9-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mf5dj\" (UID: \"1b94f190-a56d-420d-9558-7ec1116b90b9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.772146 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzz4t\" (UniqueName: \"kubernetes.io/projected/1b94f190-a56d-420d-9558-7ec1116b90b9-kube-api-access-vzz4t\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mf5dj\" (UID: \"1b94f190-a56d-420d-9558-7ec1116b90b9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.776963 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b94f190-a56d-420d-9558-7ec1116b90b9-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mf5dj\" (UID: \"1b94f190-a56d-420d-9558-7ec1116b90b9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.777314 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b94f190-a56d-420d-9558-7ec1116b90b9-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mf5dj\" (UID: \"1b94f190-a56d-420d-9558-7ec1116b90b9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.794659 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzz4t\" (UniqueName: \"kubernetes.io/projected/1b94f190-a56d-420d-9558-7ec1116b90b9-kube-api-access-vzz4t\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mf5dj\" (UID: \"1b94f190-a56d-420d-9558-7ec1116b90b9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" Oct 03 16:57:18 crc kubenswrapper[4744]: I1003 16:57:18.885392 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" Oct 03 16:57:19 crc kubenswrapper[4744]: I1003 16:57:19.491226 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj"] Oct 03 16:57:20 crc kubenswrapper[4744]: I1003 16:57:20.426055 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" event={"ID":"1b94f190-a56d-420d-9558-7ec1116b90b9","Type":"ContainerStarted","Data":"7881bbb2ef53da7a4e22e90c5e945c6108342aa1b9bff1372a3fccd32c307829"} Oct 03 16:57:21 crc kubenswrapper[4744]: I1003 16:57:21.435115 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" event={"ID":"1b94f190-a56d-420d-9558-7ec1116b90b9","Type":"ContainerStarted","Data":"178d975872978b52ae9bca94344e74d4040c304e44003e9aef092948de0e616f"} Oct 03 16:57:21 crc kubenswrapper[4744]: I1003 16:57:21.458554 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" podStartSLOduration=2.327907821 podStartE2EDuration="3.458531767s" podCreationTimestamp="2025-10-03 16:57:18 +0000 UTC" firstStartedPulling="2025-10-03 16:57:19.504563429 +0000 UTC m=+1965.784439335" lastFinishedPulling="2025-10-03 16:57:20.635187375 +0000 UTC m=+1966.915063281" observedRunningTime="2025-10-03 16:57:21.45114288 +0000 UTC m=+1967.731018796" watchObservedRunningTime="2025-10-03 16:57:21.458531767 +0000 UTC m=+1967.738407673" Oct 03 16:57:29 crc kubenswrapper[4744]: I1003 16:57:29.522548 4744 generic.go:334] "Generic (PLEG): container finished" podID="1b94f190-a56d-420d-9558-7ec1116b90b9" containerID="178d975872978b52ae9bca94344e74d4040c304e44003e9aef092948de0e616f" exitCode=0 Oct 03 16:57:29 crc kubenswrapper[4744]: I1003 16:57:29.522615 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" event={"ID":"1b94f190-a56d-420d-9558-7ec1116b90b9","Type":"ContainerDied","Data":"178d975872978b52ae9bca94344e74d4040c304e44003e9aef092948de0e616f"} Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.052126 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.168567 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzz4t\" (UniqueName: \"kubernetes.io/projected/1b94f190-a56d-420d-9558-7ec1116b90b9-kube-api-access-vzz4t\") pod \"1b94f190-a56d-420d-9558-7ec1116b90b9\" (UID: \"1b94f190-a56d-420d-9558-7ec1116b90b9\") " Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.168779 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b94f190-a56d-420d-9558-7ec1116b90b9-ssh-key\") pod \"1b94f190-a56d-420d-9558-7ec1116b90b9\" (UID: \"1b94f190-a56d-420d-9558-7ec1116b90b9\") " Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.168865 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b94f190-a56d-420d-9558-7ec1116b90b9-inventory\") pod \"1b94f190-a56d-420d-9558-7ec1116b90b9\" (UID: \"1b94f190-a56d-420d-9558-7ec1116b90b9\") " Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.178838 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b94f190-a56d-420d-9558-7ec1116b90b9-kube-api-access-vzz4t" (OuterVolumeSpecName: "kube-api-access-vzz4t") pod "1b94f190-a56d-420d-9558-7ec1116b90b9" (UID: "1b94f190-a56d-420d-9558-7ec1116b90b9"). InnerVolumeSpecName "kube-api-access-vzz4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.215038 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b94f190-a56d-420d-9558-7ec1116b90b9-inventory" (OuterVolumeSpecName: "inventory") pod "1b94f190-a56d-420d-9558-7ec1116b90b9" (UID: "1b94f190-a56d-420d-9558-7ec1116b90b9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.245050 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b94f190-a56d-420d-9558-7ec1116b90b9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1b94f190-a56d-420d-9558-7ec1116b90b9" (UID: "1b94f190-a56d-420d-9558-7ec1116b90b9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.272989 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b94f190-a56d-420d-9558-7ec1116b90b9-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.273042 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzz4t\" (UniqueName: \"kubernetes.io/projected/1b94f190-a56d-420d-9558-7ec1116b90b9-kube-api-access-vzz4t\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.273065 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b94f190-a56d-420d-9558-7ec1116b90b9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.547180 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" event={"ID":"1b94f190-a56d-420d-9558-7ec1116b90b9","Type":"ContainerDied","Data":"7881bbb2ef53da7a4e22e90c5e945c6108342aa1b9bff1372a3fccd32c307829"} Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.547708 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7881bbb2ef53da7a4e22e90c5e945c6108342aa1b9bff1372a3fccd32c307829" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.547294 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mf5dj" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.697708 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc"] Oct 03 16:57:31 crc kubenswrapper[4744]: E1003 16:57:31.698190 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b94f190-a56d-420d-9558-7ec1116b90b9" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.698213 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b94f190-a56d-420d-9558-7ec1116b90b9" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.698512 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b94f190-a56d-420d-9558-7ec1116b90b9" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.699348 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.702336 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.702600 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.703391 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.704902 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.711595 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc"] Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.782819 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pftrb\" (UniqueName: \"kubernetes.io/projected/0f862434-d64e-4c02-b8de-a120185d0053-kube-api-access-pftrb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc\" (UID: \"0f862434-d64e-4c02-b8de-a120185d0053\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.782921 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f862434-d64e-4c02-b8de-a120185d0053-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc\" (UID: \"0f862434-d64e-4c02-b8de-a120185d0053\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.783034 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f862434-d64e-4c02-b8de-a120185d0053-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc\" (UID: \"0f862434-d64e-4c02-b8de-a120185d0053\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.885363 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pftrb\" (UniqueName: \"kubernetes.io/projected/0f862434-d64e-4c02-b8de-a120185d0053-kube-api-access-pftrb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc\" (UID: \"0f862434-d64e-4c02-b8de-a120185d0053\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.885521 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f862434-d64e-4c02-b8de-a120185d0053-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc\" (UID: \"0f862434-d64e-4c02-b8de-a120185d0053\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.885630 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f862434-d64e-4c02-b8de-a120185d0053-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc\" (UID: \"0f862434-d64e-4c02-b8de-a120185d0053\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.895392 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f862434-d64e-4c02-b8de-a120185d0053-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc\" (UID: \"0f862434-d64e-4c02-b8de-a120185d0053\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.898237 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f862434-d64e-4c02-b8de-a120185d0053-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc\" (UID: \"0f862434-d64e-4c02-b8de-a120185d0053\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" Oct 03 16:57:31 crc kubenswrapper[4744]: I1003 16:57:31.919203 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pftrb\" (UniqueName: \"kubernetes.io/projected/0f862434-d64e-4c02-b8de-a120185d0053-kube-api-access-pftrb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc\" (UID: \"0f862434-d64e-4c02-b8de-a120185d0053\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" Oct 03 16:57:32 crc kubenswrapper[4744]: I1003 16:57:32.031088 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" Oct 03 16:57:32 crc kubenswrapper[4744]: I1003 16:57:32.600668 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc"] Oct 03 16:57:33 crc kubenswrapper[4744]: I1003 16:57:33.564764 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" event={"ID":"0f862434-d64e-4c02-b8de-a120185d0053","Type":"ContainerStarted","Data":"4624deda8b04711a7202d79e1ac87aad3b49ec13e8581decd2111f0d175718e6"} Oct 03 16:57:34 crc kubenswrapper[4744]: I1003 16:57:34.576802 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" event={"ID":"0f862434-d64e-4c02-b8de-a120185d0053","Type":"ContainerStarted","Data":"87ab9914a2be90a21e342e0bb97fd9140132d9c290d17f792a33add562bcd474"} Oct 03 16:57:34 crc kubenswrapper[4744]: I1003 16:57:34.602931 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" podStartSLOduration=2.590235481 podStartE2EDuration="3.602909501s" podCreationTimestamp="2025-10-03 16:57:31 +0000 UTC" firstStartedPulling="2025-10-03 16:57:32.605210539 +0000 UTC m=+1978.885086435" lastFinishedPulling="2025-10-03 16:57:33.617884549 +0000 UTC m=+1979.897760455" observedRunningTime="2025-10-03 16:57:34.598860559 +0000 UTC m=+1980.878736465" watchObservedRunningTime="2025-10-03 16:57:34.602909501 +0000 UTC m=+1980.882785417" Oct 03 16:57:39 crc kubenswrapper[4744]: I1003 16:57:39.563803 4744 scope.go:117] "RemoveContainer" containerID="147f7ab564c52abc13651319d27f4b237440be77a2aa5fc85cec734680c0a2cf" Oct 03 16:57:43 crc kubenswrapper[4744]: I1003 16:57:43.684482 4744 generic.go:334] "Generic (PLEG): container finished" podID="0f862434-d64e-4c02-b8de-a120185d0053" containerID="87ab9914a2be90a21e342e0bb97fd9140132d9c290d17f792a33add562bcd474" exitCode=0 Oct 03 16:57:43 crc kubenswrapper[4744]: I1003 16:57:43.684562 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" event={"ID":"0f862434-d64e-4c02-b8de-a120185d0053","Type":"ContainerDied","Data":"87ab9914a2be90a21e342e0bb97fd9140132d9c290d17f792a33add562bcd474"} Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.123358 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.197888 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pftrb\" (UniqueName: \"kubernetes.io/projected/0f862434-d64e-4c02-b8de-a120185d0053-kube-api-access-pftrb\") pod \"0f862434-d64e-4c02-b8de-a120185d0053\" (UID: \"0f862434-d64e-4c02-b8de-a120185d0053\") " Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.197935 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f862434-d64e-4c02-b8de-a120185d0053-ssh-key\") pod \"0f862434-d64e-4c02-b8de-a120185d0053\" (UID: \"0f862434-d64e-4c02-b8de-a120185d0053\") " Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.198049 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f862434-d64e-4c02-b8de-a120185d0053-inventory\") pod \"0f862434-d64e-4c02-b8de-a120185d0053\" (UID: \"0f862434-d64e-4c02-b8de-a120185d0053\") " Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.203288 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f862434-d64e-4c02-b8de-a120185d0053-kube-api-access-pftrb" (OuterVolumeSpecName: "kube-api-access-pftrb") pod "0f862434-d64e-4c02-b8de-a120185d0053" (UID: "0f862434-d64e-4c02-b8de-a120185d0053"). InnerVolumeSpecName "kube-api-access-pftrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.226544 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f862434-d64e-4c02-b8de-a120185d0053-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0f862434-d64e-4c02-b8de-a120185d0053" (UID: "0f862434-d64e-4c02-b8de-a120185d0053"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.227969 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f862434-d64e-4c02-b8de-a120185d0053-inventory" (OuterVolumeSpecName: "inventory") pod "0f862434-d64e-4c02-b8de-a120185d0053" (UID: "0f862434-d64e-4c02-b8de-a120185d0053"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.300318 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0f862434-d64e-4c02-b8de-a120185d0053-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.300352 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pftrb\" (UniqueName: \"kubernetes.io/projected/0f862434-d64e-4c02-b8de-a120185d0053-kube-api-access-pftrb\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.300364 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0f862434-d64e-4c02-b8de-a120185d0053-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.709867 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" event={"ID":"0f862434-d64e-4c02-b8de-a120185d0053","Type":"ContainerDied","Data":"4624deda8b04711a7202d79e1ac87aad3b49ec13e8581decd2111f0d175718e6"} Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.710186 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4624deda8b04711a7202d79e1ac87aad3b49ec13e8581decd2111f0d175718e6" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.709978 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.857806 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc"] Oct 03 16:57:45 crc kubenswrapper[4744]: E1003 16:57:45.858173 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f862434-d64e-4c02-b8de-a120185d0053" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.858188 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f862434-d64e-4c02-b8de-a120185d0053" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.858396 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f862434-d64e-4c02-b8de-a120185d0053" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.859193 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.861416 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.861486 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.861869 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.861961 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.862042 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.862566 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.864288 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.864441 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.872578 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc"] Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.909543 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m96bg\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-kube-api-access-m96bg\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.909589 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.909611 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.909667 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.909693 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.909713 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.909770 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.909791 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.909830 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.909849 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.909879 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.909928 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.909993 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:45 crc kubenswrapper[4744]: I1003 16:57:45.910015 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.011675 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.011773 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.011809 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.011876 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m96bg\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-kube-api-access-m96bg\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.011902 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.011931 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.011970 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.011995 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.012020 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.012068 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.012096 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.012181 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.012206 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.012242 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.018626 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.018750 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.019202 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.020218 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.021308 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.021918 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.021949 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.022633 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.022843 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.024189 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.024423 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.024678 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.025017 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.031843 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m96bg\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-kube-api-access-m96bg\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.196585 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:57:46 crc kubenswrapper[4744]: I1003 16:57:46.825402 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc"] Oct 03 16:57:47 crc kubenswrapper[4744]: I1003 16:57:47.743189 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" event={"ID":"3aa5f64b-0451-4034-822c-d462f01e7429","Type":"ContainerStarted","Data":"2b59cac077ce07714f95aa331be11d34a52f78d04105221be49883c7618a6933"} Oct 03 16:57:47 crc kubenswrapper[4744]: I1003 16:57:47.743798 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" event={"ID":"3aa5f64b-0451-4034-822c-d462f01e7429","Type":"ContainerStarted","Data":"fdd969900594c3bb04bb3b6cd2d54e958fe8800a41d2afd174cdb165c60a46ef"} Oct 03 16:58:26 crc kubenswrapper[4744]: I1003 16:58:26.168613 4744 generic.go:334] "Generic (PLEG): container finished" podID="3aa5f64b-0451-4034-822c-d462f01e7429" containerID="2b59cac077ce07714f95aa331be11d34a52f78d04105221be49883c7618a6933" exitCode=0 Oct 03 16:58:26 crc kubenswrapper[4744]: I1003 16:58:26.168687 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" event={"ID":"3aa5f64b-0451-4034-822c-d462f01e7429","Type":"ContainerDied","Data":"2b59cac077ce07714f95aa331be11d34a52f78d04105221be49883c7618a6933"} Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.658089 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.760899 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-bootstrap-combined-ca-bundle\") pod \"3aa5f64b-0451-4034-822c-d462f01e7429\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.761322 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-ovn-combined-ca-bundle\") pod \"3aa5f64b-0451-4034-822c-d462f01e7429\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.761379 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m96bg\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-kube-api-access-m96bg\") pod \"3aa5f64b-0451-4034-822c-d462f01e7429\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.761526 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-inventory\") pod \"3aa5f64b-0451-4034-822c-d462f01e7429\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.761597 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"3aa5f64b-0451-4034-822c-d462f01e7429\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.761722 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"3aa5f64b-0451-4034-822c-d462f01e7429\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.761779 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-nova-combined-ca-bundle\") pod \"3aa5f64b-0451-4034-822c-d462f01e7429\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.761833 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"3aa5f64b-0451-4034-822c-d462f01e7429\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.761960 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-repo-setup-combined-ca-bundle\") pod \"3aa5f64b-0451-4034-822c-d462f01e7429\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.762000 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-ovn-default-certs-0\") pod \"3aa5f64b-0451-4034-822c-d462f01e7429\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.762244 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-neutron-metadata-combined-ca-bundle\") pod \"3aa5f64b-0451-4034-822c-d462f01e7429\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.762341 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-telemetry-combined-ca-bundle\") pod \"3aa5f64b-0451-4034-822c-d462f01e7429\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.762447 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-ssh-key\") pod \"3aa5f64b-0451-4034-822c-d462f01e7429\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.762712 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-libvirt-combined-ca-bundle\") pod \"3aa5f64b-0451-4034-822c-d462f01e7429\" (UID: \"3aa5f64b-0451-4034-822c-d462f01e7429\") " Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.769565 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "3aa5f64b-0451-4034-822c-d462f01e7429" (UID: "3aa5f64b-0451-4034-822c-d462f01e7429"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.769789 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3aa5f64b-0451-4034-822c-d462f01e7429" (UID: "3aa5f64b-0451-4034-822c-d462f01e7429"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.772730 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "3aa5f64b-0451-4034-822c-d462f01e7429" (UID: "3aa5f64b-0451-4034-822c-d462f01e7429"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.777857 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "3aa5f64b-0451-4034-822c-d462f01e7429" (UID: "3aa5f64b-0451-4034-822c-d462f01e7429"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.782536 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "3aa5f64b-0451-4034-822c-d462f01e7429" (UID: "3aa5f64b-0451-4034-822c-d462f01e7429"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.782682 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "3aa5f64b-0451-4034-822c-d462f01e7429" (UID: "3aa5f64b-0451-4034-822c-d462f01e7429"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.788319 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-kube-api-access-m96bg" (OuterVolumeSpecName: "kube-api-access-m96bg") pod "3aa5f64b-0451-4034-822c-d462f01e7429" (UID: "3aa5f64b-0451-4034-822c-d462f01e7429"). InnerVolumeSpecName "kube-api-access-m96bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.788620 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "3aa5f64b-0451-4034-822c-d462f01e7429" (UID: "3aa5f64b-0451-4034-822c-d462f01e7429"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.789265 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3aa5f64b-0451-4034-822c-d462f01e7429" (UID: "3aa5f64b-0451-4034-822c-d462f01e7429"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.791970 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3aa5f64b-0451-4034-822c-d462f01e7429" (UID: "3aa5f64b-0451-4034-822c-d462f01e7429"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.792364 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "3aa5f64b-0451-4034-822c-d462f01e7429" (UID: "3aa5f64b-0451-4034-822c-d462f01e7429"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.796675 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "3aa5f64b-0451-4034-822c-d462f01e7429" (UID: "3aa5f64b-0451-4034-822c-d462f01e7429"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.823524 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-inventory" (OuterVolumeSpecName: "inventory") pod "3aa5f64b-0451-4034-822c-d462f01e7429" (UID: "3aa5f64b-0451-4034-822c-d462f01e7429"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.825827 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3aa5f64b-0451-4034-822c-d462f01e7429" (UID: "3aa5f64b-0451-4034-822c-d462f01e7429"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.869680 4744 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.869721 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.869732 4744 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.869741 4744 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.869750 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.869758 4744 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.869768 4744 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.869778 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.869787 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m96bg\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-kube-api-access-m96bg\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.869795 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.869803 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.869813 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.869822 4744 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa5f64b-0451-4034-822c-d462f01e7429-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:27 crc kubenswrapper[4744]: I1003 16:58:27.869830 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3aa5f64b-0451-4034-822c-d462f01e7429-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.193353 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" event={"ID":"3aa5f64b-0451-4034-822c-d462f01e7429","Type":"ContainerDied","Data":"fdd969900594c3bb04bb3b6cd2d54e958fe8800a41d2afd174cdb165c60a46ef"} Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.193399 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdd969900594c3bb04bb3b6cd2d54e958fe8800a41d2afd174cdb165c60a46ef" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.193477 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.330124 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2"] Oct 03 16:58:28 crc kubenswrapper[4744]: E1003 16:58:28.330883 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aa5f64b-0451-4034-822c-d462f01e7429" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.332120 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aa5f64b-0451-4034-822c-d462f01e7429" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.332624 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aa5f64b-0451-4034-822c-d462f01e7429" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.334259 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.336799 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.336873 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.337233 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.337523 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.343905 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.343917 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2"] Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.485977 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f77mn\" (UniqueName: \"kubernetes.io/projected/6c864721-c102-4044-bd81-74f3fbe29040-kube-api-access-f77mn\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8x7d2\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.486129 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8x7d2\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.486197 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8x7d2\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.486228 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8x7d2\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.486309 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6c864721-c102-4044-bd81-74f3fbe29040-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8x7d2\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.588397 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6c864721-c102-4044-bd81-74f3fbe29040-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8x7d2\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.588901 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f77mn\" (UniqueName: \"kubernetes.io/projected/6c864721-c102-4044-bd81-74f3fbe29040-kube-api-access-f77mn\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8x7d2\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.588964 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8x7d2\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.589053 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8x7d2\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.589089 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8x7d2\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.589538 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6c864721-c102-4044-bd81-74f3fbe29040-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8x7d2\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.594333 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8x7d2\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.594334 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8x7d2\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.602884 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8x7d2\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.604387 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f77mn\" (UniqueName: \"kubernetes.io/projected/6c864721-c102-4044-bd81-74f3fbe29040-kube-api-access-f77mn\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-8x7d2\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:28 crc kubenswrapper[4744]: I1003 16:58:28.678977 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:58:29 crc kubenswrapper[4744]: I1003 16:58:29.241696 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2"] Oct 03 16:58:30 crc kubenswrapper[4744]: I1003 16:58:30.212564 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" event={"ID":"6c864721-c102-4044-bd81-74f3fbe29040","Type":"ContainerStarted","Data":"932841d603005387b2f866f4cb1b3c4e4d70f2f2b91e07478f419e36974c240e"} Oct 03 16:58:30 crc kubenswrapper[4744]: I1003 16:58:30.212914 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" event={"ID":"6c864721-c102-4044-bd81-74f3fbe29040","Type":"ContainerStarted","Data":"b77bdd89bba0941aa1faa4af01b0f28c6db39cb35dada9cc6c45d620fa806ffc"} Oct 03 16:58:30 crc kubenswrapper[4744]: I1003 16:58:30.240915 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" podStartSLOduration=1.808986848 podStartE2EDuration="2.24089514s" podCreationTimestamp="2025-10-03 16:58:28 +0000 UTC" firstStartedPulling="2025-10-03 16:58:29.236454687 +0000 UTC m=+2035.516330593" lastFinishedPulling="2025-10-03 16:58:29.668362989 +0000 UTC m=+2035.948238885" observedRunningTime="2025-10-03 16:58:30.231534074 +0000 UTC m=+2036.511409970" watchObservedRunningTime="2025-10-03 16:58:30.24089514 +0000 UTC m=+2036.520771036" Oct 03 16:58:39 crc kubenswrapper[4744]: I1003 16:58:39.012929 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-84kvq"] Oct 03 16:58:39 crc kubenswrapper[4744]: I1003 16:58:39.017230 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:39 crc kubenswrapper[4744]: I1003 16:58:39.028165 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-84kvq"] Oct 03 16:58:39 crc kubenswrapper[4744]: I1003 16:58:39.140975 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f822021d-1735-4963-b87f-ddd613196106-utilities\") pod \"redhat-operators-84kvq\" (UID: \"f822021d-1735-4963-b87f-ddd613196106\") " pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:39 crc kubenswrapper[4744]: I1003 16:58:39.141413 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f822021d-1735-4963-b87f-ddd613196106-catalog-content\") pod \"redhat-operators-84kvq\" (UID: \"f822021d-1735-4963-b87f-ddd613196106\") " pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:39 crc kubenswrapper[4744]: I1003 16:58:39.141592 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs9mv\" (UniqueName: \"kubernetes.io/projected/f822021d-1735-4963-b87f-ddd613196106-kube-api-access-bs9mv\") pod \"redhat-operators-84kvq\" (UID: \"f822021d-1735-4963-b87f-ddd613196106\") " pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:39 crc kubenswrapper[4744]: I1003 16:58:39.244029 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f822021d-1735-4963-b87f-ddd613196106-utilities\") pod \"redhat-operators-84kvq\" (UID: \"f822021d-1735-4963-b87f-ddd613196106\") " pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:39 crc kubenswrapper[4744]: I1003 16:58:39.244216 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f822021d-1735-4963-b87f-ddd613196106-catalog-content\") pod \"redhat-operators-84kvq\" (UID: \"f822021d-1735-4963-b87f-ddd613196106\") " pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:39 crc kubenswrapper[4744]: I1003 16:58:39.244355 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs9mv\" (UniqueName: \"kubernetes.io/projected/f822021d-1735-4963-b87f-ddd613196106-kube-api-access-bs9mv\") pod \"redhat-operators-84kvq\" (UID: \"f822021d-1735-4963-b87f-ddd613196106\") " pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:39 crc kubenswrapper[4744]: I1003 16:58:39.244664 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f822021d-1735-4963-b87f-ddd613196106-utilities\") pod \"redhat-operators-84kvq\" (UID: \"f822021d-1735-4963-b87f-ddd613196106\") " pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:39 crc kubenswrapper[4744]: I1003 16:58:39.245088 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f822021d-1735-4963-b87f-ddd613196106-catalog-content\") pod \"redhat-operators-84kvq\" (UID: \"f822021d-1735-4963-b87f-ddd613196106\") " pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:39 crc kubenswrapper[4744]: I1003 16:58:39.274782 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs9mv\" (UniqueName: \"kubernetes.io/projected/f822021d-1735-4963-b87f-ddd613196106-kube-api-access-bs9mv\") pod \"redhat-operators-84kvq\" (UID: \"f822021d-1735-4963-b87f-ddd613196106\") " pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:39 crc kubenswrapper[4744]: I1003 16:58:39.358794 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:39 crc kubenswrapper[4744]: I1003 16:58:39.627540 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-84kvq"] Oct 03 16:58:40 crc kubenswrapper[4744]: I1003 16:58:40.342227 4744 generic.go:334] "Generic (PLEG): container finished" podID="f822021d-1735-4963-b87f-ddd613196106" containerID="7db01136ff4747832789c2bf55b023f2aeaad8d5331324eb3df16d6e250a36dc" exitCode=0 Oct 03 16:58:40 crc kubenswrapper[4744]: I1003 16:58:40.342289 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84kvq" event={"ID":"f822021d-1735-4963-b87f-ddd613196106","Type":"ContainerDied","Data":"7db01136ff4747832789c2bf55b023f2aeaad8d5331324eb3df16d6e250a36dc"} Oct 03 16:58:40 crc kubenswrapper[4744]: I1003 16:58:40.342716 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84kvq" event={"ID":"f822021d-1735-4963-b87f-ddd613196106","Type":"ContainerStarted","Data":"7c06f05a8148837cb1a3b10b76bf33ab8258a60fb1a08b0154c95009da5f2406"} Oct 03 16:58:42 crc kubenswrapper[4744]: I1003 16:58:42.378110 4744 generic.go:334] "Generic (PLEG): container finished" podID="f822021d-1735-4963-b87f-ddd613196106" containerID="f72b8ff75829ce9c7f29674900e465c730a5310f3c8089185ac6da4eeb020915" exitCode=0 Oct 03 16:58:42 crc kubenswrapper[4744]: I1003 16:58:42.378221 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84kvq" event={"ID":"f822021d-1735-4963-b87f-ddd613196106","Type":"ContainerDied","Data":"f72b8ff75829ce9c7f29674900e465c730a5310f3c8089185ac6da4eeb020915"} Oct 03 16:58:43 crc kubenswrapper[4744]: I1003 16:58:43.396821 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84kvq" event={"ID":"f822021d-1735-4963-b87f-ddd613196106","Type":"ContainerStarted","Data":"91fd5635cc446657847196313ee5e6b9413ddb96e9e038dfa95bf84d123b015e"} Oct 03 16:58:43 crc kubenswrapper[4744]: I1003 16:58:43.426263 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-84kvq" podStartSLOduration=2.825347093 podStartE2EDuration="5.42623806s" podCreationTimestamp="2025-10-03 16:58:38 +0000 UTC" firstStartedPulling="2025-10-03 16:58:40.345319067 +0000 UTC m=+2046.625194963" lastFinishedPulling="2025-10-03 16:58:42.946210034 +0000 UTC m=+2049.226085930" observedRunningTime="2025-10-03 16:58:43.416988136 +0000 UTC m=+2049.696864072" watchObservedRunningTime="2025-10-03 16:58:43.42623806 +0000 UTC m=+2049.706113986" Oct 03 16:58:49 crc kubenswrapper[4744]: I1003 16:58:49.359925 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:49 crc kubenswrapper[4744]: I1003 16:58:49.360716 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:49 crc kubenswrapper[4744]: I1003 16:58:49.445217 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:49 crc kubenswrapper[4744]: I1003 16:58:49.526874 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:49 crc kubenswrapper[4744]: I1003 16:58:49.684997 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-84kvq"] Oct 03 16:58:51 crc kubenswrapper[4744]: I1003 16:58:51.481287 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-84kvq" podUID="f822021d-1735-4963-b87f-ddd613196106" containerName="registry-server" containerID="cri-o://91fd5635cc446657847196313ee5e6b9413ddb96e9e038dfa95bf84d123b015e" gracePeriod=2 Oct 03 16:58:51 crc kubenswrapper[4744]: I1003 16:58:51.996072 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.153719 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f822021d-1735-4963-b87f-ddd613196106-catalog-content\") pod \"f822021d-1735-4963-b87f-ddd613196106\" (UID: \"f822021d-1735-4963-b87f-ddd613196106\") " Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.153793 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f822021d-1735-4963-b87f-ddd613196106-utilities\") pod \"f822021d-1735-4963-b87f-ddd613196106\" (UID: \"f822021d-1735-4963-b87f-ddd613196106\") " Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.153841 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs9mv\" (UniqueName: \"kubernetes.io/projected/f822021d-1735-4963-b87f-ddd613196106-kube-api-access-bs9mv\") pod \"f822021d-1735-4963-b87f-ddd613196106\" (UID: \"f822021d-1735-4963-b87f-ddd613196106\") " Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.155212 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f822021d-1735-4963-b87f-ddd613196106-utilities" (OuterVolumeSpecName: "utilities") pod "f822021d-1735-4963-b87f-ddd613196106" (UID: "f822021d-1735-4963-b87f-ddd613196106"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.164664 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f822021d-1735-4963-b87f-ddd613196106-kube-api-access-bs9mv" (OuterVolumeSpecName: "kube-api-access-bs9mv") pod "f822021d-1735-4963-b87f-ddd613196106" (UID: "f822021d-1735-4963-b87f-ddd613196106"). InnerVolumeSpecName "kube-api-access-bs9mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.256097 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f822021d-1735-4963-b87f-ddd613196106-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.256126 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs9mv\" (UniqueName: \"kubernetes.io/projected/f822021d-1735-4963-b87f-ddd613196106-kube-api-access-bs9mv\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.492436 4744 generic.go:334] "Generic (PLEG): container finished" podID="f822021d-1735-4963-b87f-ddd613196106" containerID="91fd5635cc446657847196313ee5e6b9413ddb96e9e038dfa95bf84d123b015e" exitCode=0 Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.492604 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84kvq" event={"ID":"f822021d-1735-4963-b87f-ddd613196106","Type":"ContainerDied","Data":"91fd5635cc446657847196313ee5e6b9413ddb96e9e038dfa95bf84d123b015e"} Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.492703 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-84kvq" event={"ID":"f822021d-1735-4963-b87f-ddd613196106","Type":"ContainerDied","Data":"7c06f05a8148837cb1a3b10b76bf33ab8258a60fb1a08b0154c95009da5f2406"} Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.492641 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-84kvq" Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.492749 4744 scope.go:117] "RemoveContainer" containerID="91fd5635cc446657847196313ee5e6b9413ddb96e9e038dfa95bf84d123b015e" Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.522095 4744 scope.go:117] "RemoveContainer" containerID="f72b8ff75829ce9c7f29674900e465c730a5310f3c8089185ac6da4eeb020915" Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.552983 4744 scope.go:117] "RemoveContainer" containerID="7db01136ff4747832789c2bf55b023f2aeaad8d5331324eb3df16d6e250a36dc" Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.602360 4744 scope.go:117] "RemoveContainer" containerID="91fd5635cc446657847196313ee5e6b9413ddb96e9e038dfa95bf84d123b015e" Oct 03 16:58:52 crc kubenswrapper[4744]: E1003 16:58:52.603171 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91fd5635cc446657847196313ee5e6b9413ddb96e9e038dfa95bf84d123b015e\": container with ID starting with 91fd5635cc446657847196313ee5e6b9413ddb96e9e038dfa95bf84d123b015e not found: ID does not exist" containerID="91fd5635cc446657847196313ee5e6b9413ddb96e9e038dfa95bf84d123b015e" Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.603444 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91fd5635cc446657847196313ee5e6b9413ddb96e9e038dfa95bf84d123b015e"} err="failed to get container status \"91fd5635cc446657847196313ee5e6b9413ddb96e9e038dfa95bf84d123b015e\": rpc error: code = NotFound desc = could not find container \"91fd5635cc446657847196313ee5e6b9413ddb96e9e038dfa95bf84d123b015e\": container with ID starting with 91fd5635cc446657847196313ee5e6b9413ddb96e9e038dfa95bf84d123b015e not found: ID does not exist" Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.603697 4744 scope.go:117] "RemoveContainer" containerID="f72b8ff75829ce9c7f29674900e465c730a5310f3c8089185ac6da4eeb020915" Oct 03 16:58:52 crc kubenswrapper[4744]: E1003 16:58:52.604330 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f72b8ff75829ce9c7f29674900e465c730a5310f3c8089185ac6da4eeb020915\": container with ID starting with f72b8ff75829ce9c7f29674900e465c730a5310f3c8089185ac6da4eeb020915 not found: ID does not exist" containerID="f72b8ff75829ce9c7f29674900e465c730a5310f3c8089185ac6da4eeb020915" Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.604657 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f72b8ff75829ce9c7f29674900e465c730a5310f3c8089185ac6da4eeb020915"} err="failed to get container status \"f72b8ff75829ce9c7f29674900e465c730a5310f3c8089185ac6da4eeb020915\": rpc error: code = NotFound desc = could not find container \"f72b8ff75829ce9c7f29674900e465c730a5310f3c8089185ac6da4eeb020915\": container with ID starting with f72b8ff75829ce9c7f29674900e465c730a5310f3c8089185ac6da4eeb020915 not found: ID does not exist" Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.604895 4744 scope.go:117] "RemoveContainer" containerID="7db01136ff4747832789c2bf55b023f2aeaad8d5331324eb3df16d6e250a36dc" Oct 03 16:58:52 crc kubenswrapper[4744]: E1003 16:58:52.605809 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7db01136ff4747832789c2bf55b023f2aeaad8d5331324eb3df16d6e250a36dc\": container with ID starting with 7db01136ff4747832789c2bf55b023f2aeaad8d5331324eb3df16d6e250a36dc not found: ID does not exist" containerID="7db01136ff4747832789c2bf55b023f2aeaad8d5331324eb3df16d6e250a36dc" Oct 03 16:58:52 crc kubenswrapper[4744]: I1003 16:58:52.605848 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7db01136ff4747832789c2bf55b023f2aeaad8d5331324eb3df16d6e250a36dc"} err="failed to get container status \"7db01136ff4747832789c2bf55b023f2aeaad8d5331324eb3df16d6e250a36dc\": rpc error: code = NotFound desc = could not find container \"7db01136ff4747832789c2bf55b023f2aeaad8d5331324eb3df16d6e250a36dc\": container with ID starting with 7db01136ff4747832789c2bf55b023f2aeaad8d5331324eb3df16d6e250a36dc not found: ID does not exist" Oct 03 16:58:53 crc kubenswrapper[4744]: I1003 16:58:53.663339 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f822021d-1735-4963-b87f-ddd613196106-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f822021d-1735-4963-b87f-ddd613196106" (UID: "f822021d-1735-4963-b87f-ddd613196106"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 16:58:53 crc kubenswrapper[4744]: I1003 16:58:53.693980 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f822021d-1735-4963-b87f-ddd613196106-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 16:58:53 crc kubenswrapper[4744]: I1003 16:58:53.749436 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-84kvq"] Oct 03 16:58:53 crc kubenswrapper[4744]: I1003 16:58:53.760048 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-84kvq"] Oct 03 16:58:54 crc kubenswrapper[4744]: I1003 16:58:54.906768 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f822021d-1735-4963-b87f-ddd613196106" path="/var/lib/kubelet/pods/f822021d-1735-4963-b87f-ddd613196106/volumes" Oct 03 16:59:04 crc kubenswrapper[4744]: I1003 16:59:04.668536 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:59:04 crc kubenswrapper[4744]: I1003 16:59:04.669185 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:59:30 crc kubenswrapper[4744]: I1003 16:59:30.939560 4744 generic.go:334] "Generic (PLEG): container finished" podID="6c864721-c102-4044-bd81-74f3fbe29040" containerID="932841d603005387b2f866f4cb1b3c4e4d70f2f2b91e07478f419e36974c240e" exitCode=0 Oct 03 16:59:30 crc kubenswrapper[4744]: I1003 16:59:30.939655 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" event={"ID":"6c864721-c102-4044-bd81-74f3fbe29040","Type":"ContainerDied","Data":"932841d603005387b2f866f4cb1b3c4e4d70f2f2b91e07478f419e36974c240e"} Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.371656 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.483963 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f77mn\" (UniqueName: \"kubernetes.io/projected/6c864721-c102-4044-bd81-74f3fbe29040-kube-api-access-f77mn\") pod \"6c864721-c102-4044-bd81-74f3fbe29040\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.484149 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6c864721-c102-4044-bd81-74f3fbe29040-ovncontroller-config-0\") pod \"6c864721-c102-4044-bd81-74f3fbe29040\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.484188 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-ovn-combined-ca-bundle\") pod \"6c864721-c102-4044-bd81-74f3fbe29040\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.484282 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-ssh-key\") pod \"6c864721-c102-4044-bd81-74f3fbe29040\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.484311 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-inventory\") pod \"6c864721-c102-4044-bd81-74f3fbe29040\" (UID: \"6c864721-c102-4044-bd81-74f3fbe29040\") " Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.497836 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "6c864721-c102-4044-bd81-74f3fbe29040" (UID: "6c864721-c102-4044-bd81-74f3fbe29040"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.497928 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c864721-c102-4044-bd81-74f3fbe29040-kube-api-access-f77mn" (OuterVolumeSpecName: "kube-api-access-f77mn") pod "6c864721-c102-4044-bd81-74f3fbe29040" (UID: "6c864721-c102-4044-bd81-74f3fbe29040"). InnerVolumeSpecName "kube-api-access-f77mn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.512926 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c864721-c102-4044-bd81-74f3fbe29040-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "6c864721-c102-4044-bd81-74f3fbe29040" (UID: "6c864721-c102-4044-bd81-74f3fbe29040"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.512926 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-inventory" (OuterVolumeSpecName: "inventory") pod "6c864721-c102-4044-bd81-74f3fbe29040" (UID: "6c864721-c102-4044-bd81-74f3fbe29040"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.519333 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6c864721-c102-4044-bd81-74f3fbe29040" (UID: "6c864721-c102-4044-bd81-74f3fbe29040"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.586856 4744 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6c864721-c102-4044-bd81-74f3fbe29040-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.586895 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.586906 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.586920 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c864721-c102-4044-bd81-74f3fbe29040-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.586931 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f77mn\" (UniqueName: \"kubernetes.io/projected/6c864721-c102-4044-bd81-74f3fbe29040-kube-api-access-f77mn\") on node \"crc\" DevicePath \"\"" Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.960527 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" event={"ID":"6c864721-c102-4044-bd81-74f3fbe29040","Type":"ContainerDied","Data":"b77bdd89bba0941aa1faa4af01b0f28c6db39cb35dada9cc6c45d620fa806ffc"} Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.960824 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b77bdd89bba0941aa1faa4af01b0f28c6db39cb35dada9cc6c45d620fa806ffc" Oct 03 16:59:32 crc kubenswrapper[4744]: I1003 16:59:32.960542 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-8x7d2" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.169319 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz"] Oct 03 16:59:33 crc kubenswrapper[4744]: E1003 16:59:33.169846 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f822021d-1735-4963-b87f-ddd613196106" containerName="extract-content" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.169865 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f822021d-1735-4963-b87f-ddd613196106" containerName="extract-content" Oct 03 16:59:33 crc kubenswrapper[4744]: E1003 16:59:33.169898 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f822021d-1735-4963-b87f-ddd613196106" containerName="extract-utilities" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.169908 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f822021d-1735-4963-b87f-ddd613196106" containerName="extract-utilities" Oct 03 16:59:33 crc kubenswrapper[4744]: E1003 16:59:33.169923 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c864721-c102-4044-bd81-74f3fbe29040" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.169932 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c864721-c102-4044-bd81-74f3fbe29040" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 16:59:33 crc kubenswrapper[4744]: E1003 16:59:33.169946 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f822021d-1735-4963-b87f-ddd613196106" containerName="registry-server" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.169953 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f822021d-1735-4963-b87f-ddd613196106" containerName="registry-server" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.170216 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f822021d-1735-4963-b87f-ddd613196106" containerName="registry-server" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.170244 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c864721-c102-4044-bd81-74f3fbe29040" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.170956 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.176068 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.176357 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.178035 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz"] Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.179643 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.179842 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.180396 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.181186 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.197842 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.197908 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.197955 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.198281 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.198669 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.199059 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8jxc\" (UniqueName: \"kubernetes.io/projected/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-kube-api-access-c8jxc\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.301004 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.301700 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.301832 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.302110 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.302256 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.302455 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8jxc\" (UniqueName: \"kubernetes.io/projected/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-kube-api-access-c8jxc\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.307198 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.307248 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.308081 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.311438 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.313983 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.320253 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8jxc\" (UniqueName: \"kubernetes.io/projected/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-kube-api-access-c8jxc\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:33 crc kubenswrapper[4744]: I1003 16:59:33.492957 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 16:59:34 crc kubenswrapper[4744]: I1003 16:59:34.117797 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz"] Oct 03 16:59:34 crc kubenswrapper[4744]: I1003 16:59:34.668855 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 16:59:34 crc kubenswrapper[4744]: I1003 16:59:34.669303 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 16:59:34 crc kubenswrapper[4744]: I1003 16:59:34.984383 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" event={"ID":"a1fb9143-c0e9-44b2-b80d-ff1ed760a115","Type":"ContainerStarted","Data":"e63e3723609c0718194929862d79b9fae68ed2da051ecb0378d89d0cd55bcdb0"} Oct 03 16:59:35 crc kubenswrapper[4744]: I1003 16:59:35.995207 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" event={"ID":"a1fb9143-c0e9-44b2-b80d-ff1ed760a115","Type":"ContainerStarted","Data":"c66b4467d12af70a333911c89def8dfd5c173cb0d41a041338f72b765a19cc75"} Oct 03 16:59:36 crc kubenswrapper[4744]: I1003 16:59:36.020153 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" podStartSLOduration=2.325681367 podStartE2EDuration="3.020131275s" podCreationTimestamp="2025-10-03 16:59:33 +0000 UTC" firstStartedPulling="2025-10-03 16:59:34.109853589 +0000 UTC m=+2100.389729505" lastFinishedPulling="2025-10-03 16:59:34.804303507 +0000 UTC m=+2101.084179413" observedRunningTime="2025-10-03 16:59:36.015111018 +0000 UTC m=+2102.294986914" watchObservedRunningTime="2025-10-03 16:59:36.020131275 +0000 UTC m=+2102.300007171" Oct 03 17:00:00 crc kubenswrapper[4744]: I1003 17:00:00.196064 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj"] Oct 03 17:00:00 crc kubenswrapper[4744]: I1003 17:00:00.197954 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" Oct 03 17:00:00 crc kubenswrapper[4744]: I1003 17:00:00.200359 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 17:00:00 crc kubenswrapper[4744]: I1003 17:00:00.200409 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 17:00:00 crc kubenswrapper[4744]: I1003 17:00:00.207997 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj"] Oct 03 17:00:00 crc kubenswrapper[4744]: I1003 17:00:00.242662 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9lvn\" (UniqueName: \"kubernetes.io/projected/8701dc0d-5341-4ed0-a151-840095189550-kube-api-access-t9lvn\") pod \"collect-profiles-29325180-8znfj\" (UID: \"8701dc0d-5341-4ed0-a151-840095189550\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" Oct 03 17:00:00 crc kubenswrapper[4744]: I1003 17:00:00.242732 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8701dc0d-5341-4ed0-a151-840095189550-config-volume\") pod \"collect-profiles-29325180-8znfj\" (UID: \"8701dc0d-5341-4ed0-a151-840095189550\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" Oct 03 17:00:00 crc kubenswrapper[4744]: I1003 17:00:00.242863 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8701dc0d-5341-4ed0-a151-840095189550-secret-volume\") pod \"collect-profiles-29325180-8znfj\" (UID: \"8701dc0d-5341-4ed0-a151-840095189550\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" Oct 03 17:00:00 crc kubenswrapper[4744]: I1003 17:00:00.344866 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8701dc0d-5341-4ed0-a151-840095189550-secret-volume\") pod \"collect-profiles-29325180-8znfj\" (UID: \"8701dc0d-5341-4ed0-a151-840095189550\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" Oct 03 17:00:00 crc kubenswrapper[4744]: I1003 17:00:00.345071 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9lvn\" (UniqueName: \"kubernetes.io/projected/8701dc0d-5341-4ed0-a151-840095189550-kube-api-access-t9lvn\") pod \"collect-profiles-29325180-8znfj\" (UID: \"8701dc0d-5341-4ed0-a151-840095189550\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" Oct 03 17:00:00 crc kubenswrapper[4744]: I1003 17:00:00.345101 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8701dc0d-5341-4ed0-a151-840095189550-config-volume\") pod \"collect-profiles-29325180-8znfj\" (UID: \"8701dc0d-5341-4ed0-a151-840095189550\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" Oct 03 17:00:00 crc kubenswrapper[4744]: I1003 17:00:00.346058 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8701dc0d-5341-4ed0-a151-840095189550-config-volume\") pod \"collect-profiles-29325180-8znfj\" (UID: \"8701dc0d-5341-4ed0-a151-840095189550\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" Oct 03 17:00:00 crc kubenswrapper[4744]: I1003 17:00:00.359446 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8701dc0d-5341-4ed0-a151-840095189550-secret-volume\") pod \"collect-profiles-29325180-8znfj\" (UID: \"8701dc0d-5341-4ed0-a151-840095189550\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" Oct 03 17:00:00 crc kubenswrapper[4744]: I1003 17:00:00.367433 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9lvn\" (UniqueName: \"kubernetes.io/projected/8701dc0d-5341-4ed0-a151-840095189550-kube-api-access-t9lvn\") pod \"collect-profiles-29325180-8znfj\" (UID: \"8701dc0d-5341-4ed0-a151-840095189550\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" Oct 03 17:00:00 crc kubenswrapper[4744]: I1003 17:00:00.520124 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" Oct 03 17:00:01 crc kubenswrapper[4744]: I1003 17:00:01.017950 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj"] Oct 03 17:00:01 crc kubenswrapper[4744]: I1003 17:00:01.248360 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" event={"ID":"8701dc0d-5341-4ed0-a151-840095189550","Type":"ContainerStarted","Data":"f7c869e728486a7ba2a9aa68269880de2a0bc033dc9f7b95e3a61cbe391f8b57"} Oct 03 17:00:02 crc kubenswrapper[4744]: I1003 17:00:02.265029 4744 generic.go:334] "Generic (PLEG): container finished" podID="8701dc0d-5341-4ed0-a151-840095189550" containerID="8f6a6cd09f3369bd5c121f96147e9edd40ef64ec9fa852ab872b20dee3985ecd" exitCode=0 Oct 03 17:00:02 crc kubenswrapper[4744]: I1003 17:00:02.265098 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" event={"ID":"8701dc0d-5341-4ed0-a151-840095189550","Type":"ContainerDied","Data":"8f6a6cd09f3369bd5c121f96147e9edd40ef64ec9fa852ab872b20dee3985ecd"} Oct 03 17:00:03 crc kubenswrapper[4744]: I1003 17:00:03.643590 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" Oct 03 17:00:03 crc kubenswrapper[4744]: I1003 17:00:03.714896 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8701dc0d-5341-4ed0-a151-840095189550-secret-volume\") pod \"8701dc0d-5341-4ed0-a151-840095189550\" (UID: \"8701dc0d-5341-4ed0-a151-840095189550\") " Oct 03 17:00:03 crc kubenswrapper[4744]: I1003 17:00:03.715007 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8701dc0d-5341-4ed0-a151-840095189550-config-volume\") pod \"8701dc0d-5341-4ed0-a151-840095189550\" (UID: \"8701dc0d-5341-4ed0-a151-840095189550\") " Oct 03 17:00:03 crc kubenswrapper[4744]: I1003 17:00:03.715215 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9lvn\" (UniqueName: \"kubernetes.io/projected/8701dc0d-5341-4ed0-a151-840095189550-kube-api-access-t9lvn\") pod \"8701dc0d-5341-4ed0-a151-840095189550\" (UID: \"8701dc0d-5341-4ed0-a151-840095189550\") " Oct 03 17:00:03 crc kubenswrapper[4744]: I1003 17:00:03.716547 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8701dc0d-5341-4ed0-a151-840095189550-config-volume" (OuterVolumeSpecName: "config-volume") pod "8701dc0d-5341-4ed0-a151-840095189550" (UID: "8701dc0d-5341-4ed0-a151-840095189550"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 17:00:03 crc kubenswrapper[4744]: I1003 17:00:03.721530 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8701dc0d-5341-4ed0-a151-840095189550-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8701dc0d-5341-4ed0-a151-840095189550" (UID: "8701dc0d-5341-4ed0-a151-840095189550"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:00:03 crc kubenswrapper[4744]: I1003 17:00:03.721928 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8701dc0d-5341-4ed0-a151-840095189550-kube-api-access-t9lvn" (OuterVolumeSpecName: "kube-api-access-t9lvn") pod "8701dc0d-5341-4ed0-a151-840095189550" (UID: "8701dc0d-5341-4ed0-a151-840095189550"). InnerVolumeSpecName "kube-api-access-t9lvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:00:03 crc kubenswrapper[4744]: I1003 17:00:03.817248 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9lvn\" (UniqueName: \"kubernetes.io/projected/8701dc0d-5341-4ed0-a151-840095189550-kube-api-access-t9lvn\") on node \"crc\" DevicePath \"\"" Oct 03 17:00:03 crc kubenswrapper[4744]: I1003 17:00:03.817284 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8701dc0d-5341-4ed0-a151-840095189550-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 17:00:03 crc kubenswrapper[4744]: I1003 17:00:03.817293 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8701dc0d-5341-4ed0-a151-840095189550-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 17:00:04 crc kubenswrapper[4744]: I1003 17:00:04.281802 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" event={"ID":"8701dc0d-5341-4ed0-a151-840095189550","Type":"ContainerDied","Data":"f7c869e728486a7ba2a9aa68269880de2a0bc033dc9f7b95e3a61cbe391f8b57"} Oct 03 17:00:04 crc kubenswrapper[4744]: I1003 17:00:04.281842 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7c869e728486a7ba2a9aa68269880de2a0bc033dc9f7b95e3a61cbe391f8b57" Oct 03 17:00:04 crc kubenswrapper[4744]: I1003 17:00:04.281933 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj" Oct 03 17:00:04 crc kubenswrapper[4744]: I1003 17:00:04.668530 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:00:04 crc kubenswrapper[4744]: I1003 17:00:04.668606 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:00:04 crc kubenswrapper[4744]: I1003 17:00:04.668664 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 17:00:04 crc kubenswrapper[4744]: I1003 17:00:04.669708 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ffa178bfa89fc417d111dc4ef6ca5fa591b9352e2e2712e9cd221a0367de2cf0"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 17:00:04 crc kubenswrapper[4744]: I1003 17:00:04.669816 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://ffa178bfa89fc417d111dc4ef6ca5fa591b9352e2e2712e9cd221a0367de2cf0" gracePeriod=600 Oct 03 17:00:04 crc kubenswrapper[4744]: I1003 17:00:04.734030 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv"] Oct 03 17:00:04 crc kubenswrapper[4744]: I1003 17:00:04.757820 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325135-jzvmv"] Oct 03 17:00:04 crc kubenswrapper[4744]: I1003 17:00:04.913052 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62497d3a-2816-455c-824e-d908c11e71e1" path="/var/lib/kubelet/pods/62497d3a-2816-455c-824e-d908c11e71e1/volumes" Oct 03 17:00:05 crc kubenswrapper[4744]: I1003 17:00:05.298343 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="ffa178bfa89fc417d111dc4ef6ca5fa591b9352e2e2712e9cd221a0367de2cf0" exitCode=0 Oct 03 17:00:05 crc kubenswrapper[4744]: I1003 17:00:05.298433 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"ffa178bfa89fc417d111dc4ef6ca5fa591b9352e2e2712e9cd221a0367de2cf0"} Oct 03 17:00:05 crc kubenswrapper[4744]: I1003 17:00:05.298517 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276"} Oct 03 17:00:05 crc kubenswrapper[4744]: I1003 17:00:05.298541 4744 scope.go:117] "RemoveContainer" containerID="4f9eecbd0f24a36474fd550d83cc8f6e42b8b1a040043d6afc3b9fc95d7f36b0" Oct 03 17:00:21 crc kubenswrapper[4744]: I1003 17:00:21.492924 4744 generic.go:334] "Generic (PLEG): container finished" podID="a1fb9143-c0e9-44b2-b80d-ff1ed760a115" containerID="c66b4467d12af70a333911c89def8dfd5c173cb0d41a041338f72b765a19cc75" exitCode=0 Oct 03 17:00:21 crc kubenswrapper[4744]: I1003 17:00:21.493032 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" event={"ID":"a1fb9143-c0e9-44b2-b80d-ff1ed760a115","Type":"ContainerDied","Data":"c66b4467d12af70a333911c89def8dfd5c173cb0d41a041338f72b765a19cc75"} Oct 03 17:00:22 crc kubenswrapper[4744]: I1003 17:00:22.975380 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.070316 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-neutron-ovn-metadata-agent-neutron-config-0\") pod \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.070679 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8jxc\" (UniqueName: \"kubernetes.io/projected/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-kube-api-access-c8jxc\") pod \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.071297 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-ssh-key\") pod \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.072111 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-nova-metadata-neutron-config-0\") pod \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.072240 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-neutron-metadata-combined-ca-bundle\") pod \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.072299 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-inventory\") pod \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\" (UID: \"a1fb9143-c0e9-44b2-b80d-ff1ed760a115\") " Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.078844 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-kube-api-access-c8jxc" (OuterVolumeSpecName: "kube-api-access-c8jxc") pod "a1fb9143-c0e9-44b2-b80d-ff1ed760a115" (UID: "a1fb9143-c0e9-44b2-b80d-ff1ed760a115"). InnerVolumeSpecName "kube-api-access-c8jxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.078898 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "a1fb9143-c0e9-44b2-b80d-ff1ed760a115" (UID: "a1fb9143-c0e9-44b2-b80d-ff1ed760a115"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.104549 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-inventory" (OuterVolumeSpecName: "inventory") pod "a1fb9143-c0e9-44b2-b80d-ff1ed760a115" (UID: "a1fb9143-c0e9-44b2-b80d-ff1ed760a115"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.121609 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a1fb9143-c0e9-44b2-b80d-ff1ed760a115" (UID: "a1fb9143-c0e9-44b2-b80d-ff1ed760a115"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.122357 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "a1fb9143-c0e9-44b2-b80d-ff1ed760a115" (UID: "a1fb9143-c0e9-44b2-b80d-ff1ed760a115"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.127241 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "a1fb9143-c0e9-44b2-b80d-ff1ed760a115" (UID: "a1fb9143-c0e9-44b2-b80d-ff1ed760a115"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.175038 4744 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.175076 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.175108 4744 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.175121 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8jxc\" (UniqueName: \"kubernetes.io/projected/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-kube-api-access-c8jxc\") on node \"crc\" DevicePath \"\"" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.175132 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.175140 4744 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a1fb9143-c0e9-44b2-b80d-ff1ed760a115-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.518404 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.520560 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz" event={"ID":"a1fb9143-c0e9-44b2-b80d-ff1ed760a115","Type":"ContainerDied","Data":"e63e3723609c0718194929862d79b9fae68ed2da051ecb0378d89d0cd55bcdb0"} Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.520633 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e63e3723609c0718194929862d79b9fae68ed2da051ecb0378d89d0cd55bcdb0" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.620032 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5"] Oct 03 17:00:23 crc kubenswrapper[4744]: E1003 17:00:23.620426 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8701dc0d-5341-4ed0-a151-840095189550" containerName="collect-profiles" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.620439 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8701dc0d-5341-4ed0-a151-840095189550" containerName="collect-profiles" Oct 03 17:00:23 crc kubenswrapper[4744]: E1003 17:00:23.620485 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1fb9143-c0e9-44b2-b80d-ff1ed760a115" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.620514 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1fb9143-c0e9-44b2-b80d-ff1ed760a115" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.620737 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1fb9143-c0e9-44b2-b80d-ff1ed760a115" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.620769 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8701dc0d-5341-4ed0-a151-840095189550" containerName="collect-profiles" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.621592 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.623405 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.624441 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.627750 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.627860 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.628010 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.630444 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5"] Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.786172 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg4hh\" (UniqueName: \"kubernetes.io/projected/78b5d479-1346-45e4-96f9-310b6b32ec50-kube-api-access-zg4hh\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.786392 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.786701 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.786894 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.786928 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.889112 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.889522 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.889587 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.889605 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.889626 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg4hh\" (UniqueName: \"kubernetes.io/projected/78b5d479-1346-45e4-96f9-310b6b32ec50-kube-api-access-zg4hh\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.896457 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.896674 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.896842 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.897087 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.907060 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg4hh\" (UniqueName: \"kubernetes.io/projected/78b5d479-1346-45e4-96f9-310b6b32ec50-kube-api-access-zg4hh\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:23 crc kubenswrapper[4744]: I1003 17:00:23.942371 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:00:24 crc kubenswrapper[4744]: I1003 17:00:24.254531 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5"] Oct 03 17:00:24 crc kubenswrapper[4744]: W1003 17:00:24.256352 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78b5d479_1346_45e4_96f9_310b6b32ec50.slice/crio-e95226a9f7e2df7acdbd2033e9233887c27f07214310aa2e96229439756775e2 WatchSource:0}: Error finding container e95226a9f7e2df7acdbd2033e9233887c27f07214310aa2e96229439756775e2: Status 404 returned error can't find the container with id e95226a9f7e2df7acdbd2033e9233887c27f07214310aa2e96229439756775e2 Oct 03 17:00:24 crc kubenswrapper[4744]: I1003 17:00:24.540421 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" event={"ID":"78b5d479-1346-45e4-96f9-310b6b32ec50","Type":"ContainerStarted","Data":"e95226a9f7e2df7acdbd2033e9233887c27f07214310aa2e96229439756775e2"} Oct 03 17:00:25 crc kubenswrapper[4744]: I1003 17:00:25.559304 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" event={"ID":"78b5d479-1346-45e4-96f9-310b6b32ec50","Type":"ContainerStarted","Data":"a351f7661ac11a8230ef89078a189e2cfd15f220c264616a6dcbcf9d49817be0"} Oct 03 17:00:25 crc kubenswrapper[4744]: I1003 17:00:25.603936 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" podStartSLOduration=2.115735506 podStartE2EDuration="2.603890456s" podCreationTimestamp="2025-10-03 17:00:23 +0000 UTC" firstStartedPulling="2025-10-03 17:00:24.258664987 +0000 UTC m=+2150.538540883" lastFinishedPulling="2025-10-03 17:00:24.746819937 +0000 UTC m=+2151.026695833" observedRunningTime="2025-10-03 17:00:25.585663446 +0000 UTC m=+2151.865539392" watchObservedRunningTime="2025-10-03 17:00:25.603890456 +0000 UTC m=+2151.883766392" Oct 03 17:00:29 crc kubenswrapper[4744]: I1003 17:00:29.862077 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-llr5p"] Oct 03 17:00:29 crc kubenswrapper[4744]: I1003 17:00:29.865146 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:29 crc kubenswrapper[4744]: I1003 17:00:29.884735 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-llr5p"] Oct 03 17:00:30 crc kubenswrapper[4744]: I1003 17:00:30.042141 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwfhp\" (UniqueName: \"kubernetes.io/projected/677a908e-01ce-4b7f-b38f-4d2a71f2d317-kube-api-access-zwfhp\") pod \"redhat-marketplace-llr5p\" (UID: \"677a908e-01ce-4b7f-b38f-4d2a71f2d317\") " pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:30 crc kubenswrapper[4744]: I1003 17:00:30.042201 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/677a908e-01ce-4b7f-b38f-4d2a71f2d317-utilities\") pod \"redhat-marketplace-llr5p\" (UID: \"677a908e-01ce-4b7f-b38f-4d2a71f2d317\") " pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:30 crc kubenswrapper[4744]: I1003 17:00:30.042456 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/677a908e-01ce-4b7f-b38f-4d2a71f2d317-catalog-content\") pod \"redhat-marketplace-llr5p\" (UID: \"677a908e-01ce-4b7f-b38f-4d2a71f2d317\") " pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:30 crc kubenswrapper[4744]: I1003 17:00:30.144336 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwfhp\" (UniqueName: \"kubernetes.io/projected/677a908e-01ce-4b7f-b38f-4d2a71f2d317-kube-api-access-zwfhp\") pod \"redhat-marketplace-llr5p\" (UID: \"677a908e-01ce-4b7f-b38f-4d2a71f2d317\") " pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:30 crc kubenswrapper[4744]: I1003 17:00:30.144378 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/677a908e-01ce-4b7f-b38f-4d2a71f2d317-utilities\") pod \"redhat-marketplace-llr5p\" (UID: \"677a908e-01ce-4b7f-b38f-4d2a71f2d317\") " pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:30 crc kubenswrapper[4744]: I1003 17:00:30.144466 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/677a908e-01ce-4b7f-b38f-4d2a71f2d317-catalog-content\") pod \"redhat-marketplace-llr5p\" (UID: \"677a908e-01ce-4b7f-b38f-4d2a71f2d317\") " pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:30 crc kubenswrapper[4744]: I1003 17:00:30.144901 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/677a908e-01ce-4b7f-b38f-4d2a71f2d317-catalog-content\") pod \"redhat-marketplace-llr5p\" (UID: \"677a908e-01ce-4b7f-b38f-4d2a71f2d317\") " pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:30 crc kubenswrapper[4744]: I1003 17:00:30.144959 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/677a908e-01ce-4b7f-b38f-4d2a71f2d317-utilities\") pod \"redhat-marketplace-llr5p\" (UID: \"677a908e-01ce-4b7f-b38f-4d2a71f2d317\") " pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:30 crc kubenswrapper[4744]: I1003 17:00:30.175422 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwfhp\" (UniqueName: \"kubernetes.io/projected/677a908e-01ce-4b7f-b38f-4d2a71f2d317-kube-api-access-zwfhp\") pod \"redhat-marketplace-llr5p\" (UID: \"677a908e-01ce-4b7f-b38f-4d2a71f2d317\") " pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:30 crc kubenswrapper[4744]: I1003 17:00:30.195779 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:30 crc kubenswrapper[4744]: I1003 17:00:30.741365 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-llr5p"] Oct 03 17:00:31 crc kubenswrapper[4744]: I1003 17:00:31.633585 4744 generic.go:334] "Generic (PLEG): container finished" podID="677a908e-01ce-4b7f-b38f-4d2a71f2d317" containerID="03ac4d9996648ce44d130afee75410646463541e2e52b0b6415236f4a09a28d4" exitCode=0 Oct 03 17:00:31 crc kubenswrapper[4744]: I1003 17:00:31.633716 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-llr5p" event={"ID":"677a908e-01ce-4b7f-b38f-4d2a71f2d317","Type":"ContainerDied","Data":"03ac4d9996648ce44d130afee75410646463541e2e52b0b6415236f4a09a28d4"} Oct 03 17:00:31 crc kubenswrapper[4744]: I1003 17:00:31.634663 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-llr5p" event={"ID":"677a908e-01ce-4b7f-b38f-4d2a71f2d317","Type":"ContainerStarted","Data":"a3c144539d5a003bd73c716000d81cfd9d5dcde2fbaaf9cbb02437e2d10a323d"} Oct 03 17:00:32 crc kubenswrapper[4744]: I1003 17:00:32.647444 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-llr5p" event={"ID":"677a908e-01ce-4b7f-b38f-4d2a71f2d317","Type":"ContainerStarted","Data":"e5e3c00de623f17315a0194221fbec5e7f2bb8527942cedbd90c4c06b07e0d45"} Oct 03 17:00:33 crc kubenswrapper[4744]: I1003 17:00:33.663745 4744 generic.go:334] "Generic (PLEG): container finished" podID="677a908e-01ce-4b7f-b38f-4d2a71f2d317" containerID="e5e3c00de623f17315a0194221fbec5e7f2bb8527942cedbd90c4c06b07e0d45" exitCode=0 Oct 03 17:00:33 crc kubenswrapper[4744]: I1003 17:00:33.663808 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-llr5p" event={"ID":"677a908e-01ce-4b7f-b38f-4d2a71f2d317","Type":"ContainerDied","Data":"e5e3c00de623f17315a0194221fbec5e7f2bb8527942cedbd90c4c06b07e0d45"} Oct 03 17:00:34 crc kubenswrapper[4744]: I1003 17:00:34.679054 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-llr5p" event={"ID":"677a908e-01ce-4b7f-b38f-4d2a71f2d317","Type":"ContainerStarted","Data":"510f1189cc0b4a0f966a30060d964641dfc8687d28191cbb2851b94870dd7d3c"} Oct 03 17:00:34 crc kubenswrapper[4744]: I1003 17:00:34.707195 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-llr5p" podStartSLOduration=3.2935856 podStartE2EDuration="5.707170429s" podCreationTimestamp="2025-10-03 17:00:29 +0000 UTC" firstStartedPulling="2025-10-03 17:00:31.638667837 +0000 UTC m=+2157.918543773" lastFinishedPulling="2025-10-03 17:00:34.052252696 +0000 UTC m=+2160.332128602" observedRunningTime="2025-10-03 17:00:34.697789337 +0000 UTC m=+2160.977665253" watchObservedRunningTime="2025-10-03 17:00:34.707170429 +0000 UTC m=+2160.987046325" Oct 03 17:00:39 crc kubenswrapper[4744]: I1003 17:00:39.762109 4744 scope.go:117] "RemoveContainer" containerID="0db7190931f983ae9e46224d23780971bab6bbee1025822a6c21787830fb70a4" Oct 03 17:00:40 crc kubenswrapper[4744]: I1003 17:00:40.196346 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:40 crc kubenswrapper[4744]: I1003 17:00:40.196919 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:40 crc kubenswrapper[4744]: I1003 17:00:40.277472 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:40 crc kubenswrapper[4744]: I1003 17:00:40.832989 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:48 crc kubenswrapper[4744]: I1003 17:00:48.793007 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-llr5p"] Oct 03 17:00:48 crc kubenswrapper[4744]: I1003 17:00:48.793996 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-llr5p" podUID="677a908e-01ce-4b7f-b38f-4d2a71f2d317" containerName="registry-server" containerID="cri-o://510f1189cc0b4a0f966a30060d964641dfc8687d28191cbb2851b94870dd7d3c" gracePeriod=2 Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.371740 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.398455 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwfhp\" (UniqueName: \"kubernetes.io/projected/677a908e-01ce-4b7f-b38f-4d2a71f2d317-kube-api-access-zwfhp\") pod \"677a908e-01ce-4b7f-b38f-4d2a71f2d317\" (UID: \"677a908e-01ce-4b7f-b38f-4d2a71f2d317\") " Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.398741 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/677a908e-01ce-4b7f-b38f-4d2a71f2d317-utilities\") pod \"677a908e-01ce-4b7f-b38f-4d2a71f2d317\" (UID: \"677a908e-01ce-4b7f-b38f-4d2a71f2d317\") " Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.398783 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/677a908e-01ce-4b7f-b38f-4d2a71f2d317-catalog-content\") pod \"677a908e-01ce-4b7f-b38f-4d2a71f2d317\" (UID: \"677a908e-01ce-4b7f-b38f-4d2a71f2d317\") " Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.401852 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/677a908e-01ce-4b7f-b38f-4d2a71f2d317-utilities" (OuterVolumeSpecName: "utilities") pod "677a908e-01ce-4b7f-b38f-4d2a71f2d317" (UID: "677a908e-01ce-4b7f-b38f-4d2a71f2d317"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.430192 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/677a908e-01ce-4b7f-b38f-4d2a71f2d317-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "677a908e-01ce-4b7f-b38f-4d2a71f2d317" (UID: "677a908e-01ce-4b7f-b38f-4d2a71f2d317"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.431825 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/677a908e-01ce-4b7f-b38f-4d2a71f2d317-kube-api-access-zwfhp" (OuterVolumeSpecName: "kube-api-access-zwfhp") pod "677a908e-01ce-4b7f-b38f-4d2a71f2d317" (UID: "677a908e-01ce-4b7f-b38f-4d2a71f2d317"). InnerVolumeSpecName "kube-api-access-zwfhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.501810 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/677a908e-01ce-4b7f-b38f-4d2a71f2d317-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.501840 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/677a908e-01ce-4b7f-b38f-4d2a71f2d317-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.501850 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwfhp\" (UniqueName: \"kubernetes.io/projected/677a908e-01ce-4b7f-b38f-4d2a71f2d317-kube-api-access-zwfhp\") on node \"crc\" DevicePath \"\"" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.856915 4744 generic.go:334] "Generic (PLEG): container finished" podID="677a908e-01ce-4b7f-b38f-4d2a71f2d317" containerID="510f1189cc0b4a0f966a30060d964641dfc8687d28191cbb2851b94870dd7d3c" exitCode=0 Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.856959 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-llr5p" event={"ID":"677a908e-01ce-4b7f-b38f-4d2a71f2d317","Type":"ContainerDied","Data":"510f1189cc0b4a0f966a30060d964641dfc8687d28191cbb2851b94870dd7d3c"} Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.856983 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-llr5p" event={"ID":"677a908e-01ce-4b7f-b38f-4d2a71f2d317","Type":"ContainerDied","Data":"a3c144539d5a003bd73c716000d81cfd9d5dcde2fbaaf9cbb02437e2d10a323d"} Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.856985 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-llr5p" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.857052 4744 scope.go:117] "RemoveContainer" containerID="510f1189cc0b4a0f966a30060d964641dfc8687d28191cbb2851b94870dd7d3c" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.877404 4744 scope.go:117] "RemoveContainer" containerID="e5e3c00de623f17315a0194221fbec5e7f2bb8527942cedbd90c4c06b07e0d45" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.896299 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-llr5p"] Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.907051 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-llr5p"] Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.923125 4744 scope.go:117] "RemoveContainer" containerID="03ac4d9996648ce44d130afee75410646463541e2e52b0b6415236f4a09a28d4" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.972666 4744 scope.go:117] "RemoveContainer" containerID="510f1189cc0b4a0f966a30060d964641dfc8687d28191cbb2851b94870dd7d3c" Oct 03 17:00:49 crc kubenswrapper[4744]: E1003 17:00:49.974666 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"510f1189cc0b4a0f966a30060d964641dfc8687d28191cbb2851b94870dd7d3c\": container with ID starting with 510f1189cc0b4a0f966a30060d964641dfc8687d28191cbb2851b94870dd7d3c not found: ID does not exist" containerID="510f1189cc0b4a0f966a30060d964641dfc8687d28191cbb2851b94870dd7d3c" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.974745 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"510f1189cc0b4a0f966a30060d964641dfc8687d28191cbb2851b94870dd7d3c"} err="failed to get container status \"510f1189cc0b4a0f966a30060d964641dfc8687d28191cbb2851b94870dd7d3c\": rpc error: code = NotFound desc = could not find container \"510f1189cc0b4a0f966a30060d964641dfc8687d28191cbb2851b94870dd7d3c\": container with ID starting with 510f1189cc0b4a0f966a30060d964641dfc8687d28191cbb2851b94870dd7d3c not found: ID does not exist" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.974786 4744 scope.go:117] "RemoveContainer" containerID="e5e3c00de623f17315a0194221fbec5e7f2bb8527942cedbd90c4c06b07e0d45" Oct 03 17:00:49 crc kubenswrapper[4744]: E1003 17:00:49.975150 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5e3c00de623f17315a0194221fbec5e7f2bb8527942cedbd90c4c06b07e0d45\": container with ID starting with e5e3c00de623f17315a0194221fbec5e7f2bb8527942cedbd90c4c06b07e0d45 not found: ID does not exist" containerID="e5e3c00de623f17315a0194221fbec5e7f2bb8527942cedbd90c4c06b07e0d45" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.975190 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5e3c00de623f17315a0194221fbec5e7f2bb8527942cedbd90c4c06b07e0d45"} err="failed to get container status \"e5e3c00de623f17315a0194221fbec5e7f2bb8527942cedbd90c4c06b07e0d45\": rpc error: code = NotFound desc = could not find container \"e5e3c00de623f17315a0194221fbec5e7f2bb8527942cedbd90c4c06b07e0d45\": container with ID starting with e5e3c00de623f17315a0194221fbec5e7f2bb8527942cedbd90c4c06b07e0d45 not found: ID does not exist" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.975221 4744 scope.go:117] "RemoveContainer" containerID="03ac4d9996648ce44d130afee75410646463541e2e52b0b6415236f4a09a28d4" Oct 03 17:00:49 crc kubenswrapper[4744]: E1003 17:00:49.975590 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03ac4d9996648ce44d130afee75410646463541e2e52b0b6415236f4a09a28d4\": container with ID starting with 03ac4d9996648ce44d130afee75410646463541e2e52b0b6415236f4a09a28d4 not found: ID does not exist" containerID="03ac4d9996648ce44d130afee75410646463541e2e52b0b6415236f4a09a28d4" Oct 03 17:00:49 crc kubenswrapper[4744]: I1003 17:00:49.975662 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03ac4d9996648ce44d130afee75410646463541e2e52b0b6415236f4a09a28d4"} err="failed to get container status \"03ac4d9996648ce44d130afee75410646463541e2e52b0b6415236f4a09a28d4\": rpc error: code = NotFound desc = could not find container \"03ac4d9996648ce44d130afee75410646463541e2e52b0b6415236f4a09a28d4\": container with ID starting with 03ac4d9996648ce44d130afee75410646463541e2e52b0b6415236f4a09a28d4 not found: ID does not exist" Oct 03 17:00:50 crc kubenswrapper[4744]: I1003 17:00:50.903412 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="677a908e-01ce-4b7f-b38f-4d2a71f2d317" path="/var/lib/kubelet/pods/677a908e-01ce-4b7f-b38f-4d2a71f2d317/volumes" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.160966 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29325181-pc2dm"] Oct 03 17:01:00 crc kubenswrapper[4744]: E1003 17:01:00.162197 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="677a908e-01ce-4b7f-b38f-4d2a71f2d317" containerName="extract-content" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.162219 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="677a908e-01ce-4b7f-b38f-4d2a71f2d317" containerName="extract-content" Oct 03 17:01:00 crc kubenswrapper[4744]: E1003 17:01:00.162261 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="677a908e-01ce-4b7f-b38f-4d2a71f2d317" containerName="extract-utilities" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.162271 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="677a908e-01ce-4b7f-b38f-4d2a71f2d317" containerName="extract-utilities" Oct 03 17:01:00 crc kubenswrapper[4744]: E1003 17:01:00.162296 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="677a908e-01ce-4b7f-b38f-4d2a71f2d317" containerName="registry-server" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.162306 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="677a908e-01ce-4b7f-b38f-4d2a71f2d317" containerName="registry-server" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.162637 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="677a908e-01ce-4b7f-b38f-4d2a71f2d317" containerName="registry-server" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.163649 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.174678 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325181-pc2dm"] Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.245439 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rffjp\" (UniqueName: \"kubernetes.io/projected/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-kube-api-access-rffjp\") pod \"keystone-cron-29325181-pc2dm\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.245634 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-fernet-keys\") pod \"keystone-cron-29325181-pc2dm\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.245738 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-combined-ca-bundle\") pod \"keystone-cron-29325181-pc2dm\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.245794 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-config-data\") pod \"keystone-cron-29325181-pc2dm\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.347572 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-combined-ca-bundle\") pod \"keystone-cron-29325181-pc2dm\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.347685 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-config-data\") pod \"keystone-cron-29325181-pc2dm\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.347935 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rffjp\" (UniqueName: \"kubernetes.io/projected/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-kube-api-access-rffjp\") pod \"keystone-cron-29325181-pc2dm\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.348133 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-fernet-keys\") pod \"keystone-cron-29325181-pc2dm\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.353915 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-fernet-keys\") pod \"keystone-cron-29325181-pc2dm\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.354479 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-config-data\") pod \"keystone-cron-29325181-pc2dm\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.362290 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-combined-ca-bundle\") pod \"keystone-cron-29325181-pc2dm\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.383050 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rffjp\" (UniqueName: \"kubernetes.io/projected/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-kube-api-access-rffjp\") pod \"keystone-cron-29325181-pc2dm\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.486817 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.958398 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325181-pc2dm"] Oct 03 17:01:00 crc kubenswrapper[4744]: I1003 17:01:00.983810 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325181-pc2dm" event={"ID":"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6","Type":"ContainerStarted","Data":"6958c560e58e87e927cb55ae562b61d240106b824eb329237a484f08cd645c23"} Oct 03 17:01:01 crc kubenswrapper[4744]: I1003 17:01:01.996420 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325181-pc2dm" event={"ID":"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6","Type":"ContainerStarted","Data":"86999d359b38a96e9ab268995a9526489ff3f51586a53fe25922205b2db566d7"} Oct 03 17:01:02 crc kubenswrapper[4744]: I1003 17:01:02.017224 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29325181-pc2dm" podStartSLOduration=2.017200801 podStartE2EDuration="2.017200801s" podCreationTimestamp="2025-10-03 17:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 17:01:02.016708499 +0000 UTC m=+2188.296584405" watchObservedRunningTime="2025-10-03 17:01:02.017200801 +0000 UTC m=+2188.297076707" Oct 03 17:01:04 crc kubenswrapper[4744]: I1003 17:01:04.015188 4744 generic.go:334] "Generic (PLEG): container finished" podID="513c18a1-42b9-43f1-bb0c-4a99a3c45ea6" containerID="86999d359b38a96e9ab268995a9526489ff3f51586a53fe25922205b2db566d7" exitCode=0 Oct 03 17:01:04 crc kubenswrapper[4744]: I1003 17:01:04.015272 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325181-pc2dm" event={"ID":"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6","Type":"ContainerDied","Data":"86999d359b38a96e9ab268995a9526489ff3f51586a53fe25922205b2db566d7"} Oct 03 17:01:05 crc kubenswrapper[4744]: I1003 17:01:05.451611 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:01:05 crc kubenswrapper[4744]: I1003 17:01:05.479156 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-config-data\") pod \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " Oct 03 17:01:05 crc kubenswrapper[4744]: I1003 17:01:05.479260 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-fernet-keys\") pod \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " Oct 03 17:01:05 crc kubenswrapper[4744]: I1003 17:01:05.479317 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rffjp\" (UniqueName: \"kubernetes.io/projected/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-kube-api-access-rffjp\") pod \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " Oct 03 17:01:05 crc kubenswrapper[4744]: I1003 17:01:05.479440 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-combined-ca-bundle\") pod \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\" (UID: \"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6\") " Oct 03 17:01:05 crc kubenswrapper[4744]: I1003 17:01:05.524024 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-kube-api-access-rffjp" (OuterVolumeSpecName: "kube-api-access-rffjp") pod "513c18a1-42b9-43f1-bb0c-4a99a3c45ea6" (UID: "513c18a1-42b9-43f1-bb0c-4a99a3c45ea6"). InnerVolumeSpecName "kube-api-access-rffjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:01:05 crc kubenswrapper[4744]: I1003 17:01:05.525464 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "513c18a1-42b9-43f1-bb0c-4a99a3c45ea6" (UID: "513c18a1-42b9-43f1-bb0c-4a99a3c45ea6"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:01:05 crc kubenswrapper[4744]: I1003 17:01:05.531060 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "513c18a1-42b9-43f1-bb0c-4a99a3c45ea6" (UID: "513c18a1-42b9-43f1-bb0c-4a99a3c45ea6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:01:05 crc kubenswrapper[4744]: I1003 17:01:05.560593 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-config-data" (OuterVolumeSpecName: "config-data") pod "513c18a1-42b9-43f1-bb0c-4a99a3c45ea6" (UID: "513c18a1-42b9-43f1-bb0c-4a99a3c45ea6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:01:05 crc kubenswrapper[4744]: I1003 17:01:05.582292 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 17:01:05 crc kubenswrapper[4744]: I1003 17:01:05.582328 4744 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 03 17:01:05 crc kubenswrapper[4744]: I1003 17:01:05.582341 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rffjp\" (UniqueName: \"kubernetes.io/projected/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-kube-api-access-rffjp\") on node \"crc\" DevicePath \"\"" Oct 03 17:01:05 crc kubenswrapper[4744]: I1003 17:01:05.582356 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513c18a1-42b9-43f1-bb0c-4a99a3c45ea6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 17:01:06 crc kubenswrapper[4744]: I1003 17:01:06.039630 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325181-pc2dm" event={"ID":"513c18a1-42b9-43f1-bb0c-4a99a3c45ea6","Type":"ContainerDied","Data":"6958c560e58e87e927cb55ae562b61d240106b824eb329237a484f08cd645c23"} Oct 03 17:01:06 crc kubenswrapper[4744]: I1003 17:01:06.039671 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6958c560e58e87e927cb55ae562b61d240106b824eb329237a484f08cd645c23" Oct 03 17:01:06 crc kubenswrapper[4744]: I1003 17:01:06.039695 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325181-pc2dm" Oct 03 17:02:34 crc kubenswrapper[4744]: I1003 17:02:34.668318 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:02:34 crc kubenswrapper[4744]: I1003 17:02:34.669037 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:03:04 crc kubenswrapper[4744]: I1003 17:03:04.668487 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:03:04 crc kubenswrapper[4744]: I1003 17:03:04.669219 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:03:34 crc kubenswrapper[4744]: I1003 17:03:34.668289 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:03:34 crc kubenswrapper[4744]: I1003 17:03:34.668803 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:03:34 crc kubenswrapper[4744]: I1003 17:03:34.668848 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 17:03:34 crc kubenswrapper[4744]: I1003 17:03:34.669659 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 17:03:34 crc kubenswrapper[4744]: I1003 17:03:34.669705 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" gracePeriod=600 Oct 03 17:03:34 crc kubenswrapper[4744]: E1003 17:03:34.818406 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:03:35 crc kubenswrapper[4744]: I1003 17:03:35.719206 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" exitCode=0 Oct 03 17:03:35 crc kubenswrapper[4744]: I1003 17:03:35.719313 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276"} Oct 03 17:03:35 crc kubenswrapper[4744]: I1003 17:03:35.719775 4744 scope.go:117] "RemoveContainer" containerID="ffa178bfa89fc417d111dc4ef6ca5fa591b9352e2e2712e9cd221a0367de2cf0" Oct 03 17:03:35 crc kubenswrapper[4744]: I1003 17:03:35.721258 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:03:35 crc kubenswrapper[4744]: E1003 17:03:35.721833 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:03:50 crc kubenswrapper[4744]: I1003 17:03:50.906364 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:03:50 crc kubenswrapper[4744]: E1003 17:03:50.907896 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:04:01 crc kubenswrapper[4744]: I1003 17:04:01.891943 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:04:01 crc kubenswrapper[4744]: E1003 17:04:01.892725 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:04:16 crc kubenswrapper[4744]: I1003 17:04:16.892449 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:04:16 crc kubenswrapper[4744]: E1003 17:04:16.893861 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:04:21 crc kubenswrapper[4744]: I1003 17:04:21.256402 4744 generic.go:334] "Generic (PLEG): container finished" podID="78b5d479-1346-45e4-96f9-310b6b32ec50" containerID="a351f7661ac11a8230ef89078a189e2cfd15f220c264616a6dcbcf9d49817be0" exitCode=0 Oct 03 17:04:21 crc kubenswrapper[4744]: I1003 17:04:21.256545 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" event={"ID":"78b5d479-1346-45e4-96f9-310b6b32ec50","Type":"ContainerDied","Data":"a351f7661ac11a8230ef89078a189e2cfd15f220c264616a6dcbcf9d49817be0"} Oct 03 17:04:22 crc kubenswrapper[4744]: I1003 17:04:22.792093 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:04:22 crc kubenswrapper[4744]: I1003 17:04:22.924301 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-libvirt-combined-ca-bundle\") pod \"78b5d479-1346-45e4-96f9-310b6b32ec50\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " Oct 03 17:04:22 crc kubenswrapper[4744]: I1003 17:04:22.924381 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg4hh\" (UniqueName: \"kubernetes.io/projected/78b5d479-1346-45e4-96f9-310b6b32ec50-kube-api-access-zg4hh\") pod \"78b5d479-1346-45e4-96f9-310b6b32ec50\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " Oct 03 17:04:22 crc kubenswrapper[4744]: I1003 17:04:22.924432 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-inventory\") pod \"78b5d479-1346-45e4-96f9-310b6b32ec50\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " Oct 03 17:04:22 crc kubenswrapper[4744]: I1003 17:04:22.924515 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-ssh-key\") pod \"78b5d479-1346-45e4-96f9-310b6b32ec50\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " Oct 03 17:04:22 crc kubenswrapper[4744]: I1003 17:04:22.924583 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-libvirt-secret-0\") pod \"78b5d479-1346-45e4-96f9-310b6b32ec50\" (UID: \"78b5d479-1346-45e4-96f9-310b6b32ec50\") " Oct 03 17:04:22 crc kubenswrapper[4744]: I1003 17:04:22.942760 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "78b5d479-1346-45e4-96f9-310b6b32ec50" (UID: "78b5d479-1346-45e4-96f9-310b6b32ec50"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:04:22 crc kubenswrapper[4744]: I1003 17:04:22.958210 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "78b5d479-1346-45e4-96f9-310b6b32ec50" (UID: "78b5d479-1346-45e4-96f9-310b6b32ec50"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:04:22 crc kubenswrapper[4744]: I1003 17:04:22.963736 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78b5d479-1346-45e4-96f9-310b6b32ec50-kube-api-access-zg4hh" (OuterVolumeSpecName: "kube-api-access-zg4hh") pod "78b5d479-1346-45e4-96f9-310b6b32ec50" (UID: "78b5d479-1346-45e4-96f9-310b6b32ec50"). InnerVolumeSpecName "kube-api-access-zg4hh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:04:22 crc kubenswrapper[4744]: I1003 17:04:22.966289 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "78b5d479-1346-45e4-96f9-310b6b32ec50" (UID: "78b5d479-1346-45e4-96f9-310b6b32ec50"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:04:22 crc kubenswrapper[4744]: I1003 17:04:22.981964 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-inventory" (OuterVolumeSpecName: "inventory") pod "78b5d479-1346-45e4-96f9-310b6b32ec50" (UID: "78b5d479-1346-45e4-96f9-310b6b32ec50"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.030799 4744 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.030843 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zg4hh\" (UniqueName: \"kubernetes.io/projected/78b5d479-1346-45e4-96f9-310b6b32ec50-kube-api-access-zg4hh\") on node \"crc\" DevicePath \"\"" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.030856 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.030871 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.030883 4744 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/78b5d479-1346-45e4-96f9-310b6b32ec50-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.286243 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" event={"ID":"78b5d479-1346-45e4-96f9-310b6b32ec50","Type":"ContainerDied","Data":"e95226a9f7e2df7acdbd2033e9233887c27f07214310aa2e96229439756775e2"} Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.286302 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e95226a9f7e2df7acdbd2033e9233887c27f07214310aa2e96229439756775e2" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.286413 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.452898 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b"] Oct 03 17:04:23 crc kubenswrapper[4744]: E1003 17:04:23.453427 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="513c18a1-42b9-43f1-bb0c-4a99a3c45ea6" containerName="keystone-cron" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.453454 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="513c18a1-42b9-43f1-bb0c-4a99a3c45ea6" containerName="keystone-cron" Oct 03 17:04:23 crc kubenswrapper[4744]: E1003 17:04:23.453489 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78b5d479-1346-45e4-96f9-310b6b32ec50" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.453524 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="78b5d479-1346-45e4-96f9-310b6b32ec50" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.453821 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="513c18a1-42b9-43f1-bb0c-4a99a3c45ea6" containerName="keystone-cron" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.453870 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="78b5d479-1346-45e4-96f9-310b6b32ec50" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.454664 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.457070 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.457533 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.457877 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.458478 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.458487 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.458839 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.461007 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.554818 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.555239 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.555335 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cwfm\" (UniqueName: \"kubernetes.io/projected/710d1522-5c50-4e85-bb48-a099cfbb892d-kube-api-access-6cwfm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.555379 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.555415 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.555530 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.555590 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.555651 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.555759 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.559909 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b"] Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.657273 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.657344 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.657380 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.657423 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cwfm\" (UniqueName: \"kubernetes.io/projected/710d1522-5c50-4e85-bb48-a099cfbb892d-kube-api-access-6cwfm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.657446 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.657465 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.657531 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.657565 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.657595 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.659090 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.662479 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.662824 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.663230 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.663740 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.665011 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.667397 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.669365 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.676686 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cwfm\" (UniqueName: \"kubernetes.io/projected/710d1522-5c50-4e85-bb48-a099cfbb892d-kube-api-access-6cwfm\") pod \"nova-edpm-deployment-openstack-edpm-ipam-z279b\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:23 crc kubenswrapper[4744]: I1003 17:04:23.776198 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:04:24 crc kubenswrapper[4744]: I1003 17:04:24.392288 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b"] Oct 03 17:04:24 crc kubenswrapper[4744]: I1003 17:04:24.397917 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 17:04:25 crc kubenswrapper[4744]: I1003 17:04:25.306672 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" event={"ID":"710d1522-5c50-4e85-bb48-a099cfbb892d","Type":"ContainerStarted","Data":"c69e53eb25e5bd2e5b0ca33902739a78ab7d25e4ef793c4f2f6c8ee9e67a15c0"} Oct 03 17:04:25 crc kubenswrapper[4744]: I1003 17:04:25.307099 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" event={"ID":"710d1522-5c50-4e85-bb48-a099cfbb892d","Type":"ContainerStarted","Data":"76d2040728269f6e9b638c488e4d7fcbfd678b4fbcbdcca16934641e4226da7f"} Oct 03 17:04:25 crc kubenswrapper[4744]: I1003 17:04:25.331285 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" podStartSLOduration=1.858760734 podStartE2EDuration="2.331262298s" podCreationTimestamp="2025-10-03 17:04:23 +0000 UTC" firstStartedPulling="2025-10-03 17:04:24.397441626 +0000 UTC m=+2390.677317562" lastFinishedPulling="2025-10-03 17:04:24.86994323 +0000 UTC m=+2391.149819126" observedRunningTime="2025-10-03 17:04:25.327137916 +0000 UTC m=+2391.607013832" watchObservedRunningTime="2025-10-03 17:04:25.331262298 +0000 UTC m=+2391.611138204" Oct 03 17:04:31 crc kubenswrapper[4744]: I1003 17:04:31.891994 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:04:31 crc kubenswrapper[4744]: E1003 17:04:31.892749 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:04:42 crc kubenswrapper[4744]: I1003 17:04:42.892550 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:04:42 crc kubenswrapper[4744]: E1003 17:04:42.893485 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:04:54 crc kubenswrapper[4744]: I1003 17:04:54.905589 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:04:54 crc kubenswrapper[4744]: E1003 17:04:54.907393 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:05:06 crc kubenswrapper[4744]: I1003 17:05:06.892634 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:05:06 crc kubenswrapper[4744]: E1003 17:05:06.893732 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:05:17 crc kubenswrapper[4744]: I1003 17:05:17.891797 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:05:17 crc kubenswrapper[4744]: E1003 17:05:17.892600 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:05:31 crc kubenswrapper[4744]: I1003 17:05:31.892761 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:05:31 crc kubenswrapper[4744]: E1003 17:05:31.893635 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:05:45 crc kubenswrapper[4744]: I1003 17:05:45.894152 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:05:45 crc kubenswrapper[4744]: E1003 17:05:45.895446 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:05:59 crc kubenswrapper[4744]: I1003 17:05:59.891842 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:05:59 crc kubenswrapper[4744]: E1003 17:05:59.893036 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:06:12 crc kubenswrapper[4744]: I1003 17:06:12.893852 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:06:12 crc kubenswrapper[4744]: E1003 17:06:12.895175 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:06:25 crc kubenswrapper[4744]: I1003 17:06:25.892127 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:06:25 crc kubenswrapper[4744]: E1003 17:06:25.892861 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:06:38 crc kubenswrapper[4744]: I1003 17:06:38.891997 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:06:38 crc kubenswrapper[4744]: E1003 17:06:38.893953 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:06:50 crc kubenswrapper[4744]: I1003 17:06:50.892464 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:06:50 crc kubenswrapper[4744]: E1003 17:06:50.893324 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:07:03 crc kubenswrapper[4744]: I1003 17:07:03.892806 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:07:03 crc kubenswrapper[4744]: E1003 17:07:03.894077 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:07:17 crc kubenswrapper[4744]: I1003 17:07:17.891960 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:07:17 crc kubenswrapper[4744]: E1003 17:07:17.892925 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:07:28 crc kubenswrapper[4744]: I1003 17:07:28.893012 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:07:28 crc kubenswrapper[4744]: E1003 17:07:28.894170 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:07:31 crc kubenswrapper[4744]: I1003 17:07:31.570381 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lx9fw"] Oct 03 17:07:31 crc kubenswrapper[4744]: I1003 17:07:31.575837 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:31 crc kubenswrapper[4744]: I1003 17:07:31.586914 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lx9fw"] Oct 03 17:07:31 crc kubenswrapper[4744]: I1003 17:07:31.644563 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl8r5\" (UniqueName: \"kubernetes.io/projected/5209fede-d923-45ff-9f9e-b17962c687b2-kube-api-access-jl8r5\") pod \"certified-operators-lx9fw\" (UID: \"5209fede-d923-45ff-9f9e-b17962c687b2\") " pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:31 crc kubenswrapper[4744]: I1003 17:07:31.644980 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5209fede-d923-45ff-9f9e-b17962c687b2-catalog-content\") pod \"certified-operators-lx9fw\" (UID: \"5209fede-d923-45ff-9f9e-b17962c687b2\") " pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:31 crc kubenswrapper[4744]: I1003 17:07:31.645389 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5209fede-d923-45ff-9f9e-b17962c687b2-utilities\") pod \"certified-operators-lx9fw\" (UID: \"5209fede-d923-45ff-9f9e-b17962c687b2\") " pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:31 crc kubenswrapper[4744]: I1003 17:07:31.747551 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5209fede-d923-45ff-9f9e-b17962c687b2-utilities\") pod \"certified-operators-lx9fw\" (UID: \"5209fede-d923-45ff-9f9e-b17962c687b2\") " pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:31 crc kubenswrapper[4744]: I1003 17:07:31.747668 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl8r5\" (UniqueName: \"kubernetes.io/projected/5209fede-d923-45ff-9f9e-b17962c687b2-kube-api-access-jl8r5\") pod \"certified-operators-lx9fw\" (UID: \"5209fede-d923-45ff-9f9e-b17962c687b2\") " pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:31 crc kubenswrapper[4744]: I1003 17:07:31.747884 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5209fede-d923-45ff-9f9e-b17962c687b2-catalog-content\") pod \"certified-operators-lx9fw\" (UID: \"5209fede-d923-45ff-9f9e-b17962c687b2\") " pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:31 crc kubenswrapper[4744]: I1003 17:07:31.748350 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5209fede-d923-45ff-9f9e-b17962c687b2-utilities\") pod \"certified-operators-lx9fw\" (UID: \"5209fede-d923-45ff-9f9e-b17962c687b2\") " pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:31 crc kubenswrapper[4744]: I1003 17:07:31.748794 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5209fede-d923-45ff-9f9e-b17962c687b2-catalog-content\") pod \"certified-operators-lx9fw\" (UID: \"5209fede-d923-45ff-9f9e-b17962c687b2\") " pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:31 crc kubenswrapper[4744]: I1003 17:07:31.774325 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl8r5\" (UniqueName: \"kubernetes.io/projected/5209fede-d923-45ff-9f9e-b17962c687b2-kube-api-access-jl8r5\") pod \"certified-operators-lx9fw\" (UID: \"5209fede-d923-45ff-9f9e-b17962c687b2\") " pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:31 crc kubenswrapper[4744]: I1003 17:07:31.923961 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:32 crc kubenswrapper[4744]: I1003 17:07:32.468292 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lx9fw"] Oct 03 17:07:33 crc kubenswrapper[4744]: I1003 17:07:33.457469 4744 generic.go:334] "Generic (PLEG): container finished" podID="710d1522-5c50-4e85-bb48-a099cfbb892d" containerID="c69e53eb25e5bd2e5b0ca33902739a78ab7d25e4ef793c4f2f6c8ee9e67a15c0" exitCode=0 Oct 03 17:07:33 crc kubenswrapper[4744]: I1003 17:07:33.457943 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" event={"ID":"710d1522-5c50-4e85-bb48-a099cfbb892d","Type":"ContainerDied","Data":"c69e53eb25e5bd2e5b0ca33902739a78ab7d25e4ef793c4f2f6c8ee9e67a15c0"} Oct 03 17:07:33 crc kubenswrapper[4744]: I1003 17:07:33.461272 4744 generic.go:334] "Generic (PLEG): container finished" podID="5209fede-d923-45ff-9f9e-b17962c687b2" containerID="9aa4ff0e2b9dd21b80a626f3813971cb1d5810f18c44eb5f1b07f0a80bd5acfb" exitCode=0 Oct 03 17:07:33 crc kubenswrapper[4744]: I1003 17:07:33.461350 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx9fw" event={"ID":"5209fede-d923-45ff-9f9e-b17962c687b2","Type":"ContainerDied","Data":"9aa4ff0e2b9dd21b80a626f3813971cb1d5810f18c44eb5f1b07f0a80bd5acfb"} Oct 03 17:07:33 crc kubenswrapper[4744]: I1003 17:07:33.461393 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx9fw" event={"ID":"5209fede-d923-45ff-9f9e-b17962c687b2","Type":"ContainerStarted","Data":"1013dcc75d4e2a81a6c9b1ea74135f3547706c554d9ed6769752ac7cbbab4a4e"} Oct 03 17:07:34 crc kubenswrapper[4744]: I1003 17:07:34.973556 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.048940 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-cell1-compute-config-0\") pod \"710d1522-5c50-4e85-bb48-a099cfbb892d\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.049008 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cwfm\" (UniqueName: \"kubernetes.io/projected/710d1522-5c50-4e85-bb48-a099cfbb892d-kube-api-access-6cwfm\") pod \"710d1522-5c50-4e85-bb48-a099cfbb892d\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.049059 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-combined-ca-bundle\") pod \"710d1522-5c50-4e85-bb48-a099cfbb892d\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.049093 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-cell1-compute-config-1\") pod \"710d1522-5c50-4e85-bb48-a099cfbb892d\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.049158 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-migration-ssh-key-0\") pod \"710d1522-5c50-4e85-bb48-a099cfbb892d\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.049181 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-migration-ssh-key-1\") pod \"710d1522-5c50-4e85-bb48-a099cfbb892d\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.049303 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-ssh-key\") pod \"710d1522-5c50-4e85-bb48-a099cfbb892d\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.049342 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-extra-config-0\") pod \"710d1522-5c50-4e85-bb48-a099cfbb892d\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.049373 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-inventory\") pod \"710d1522-5c50-4e85-bb48-a099cfbb892d\" (UID: \"710d1522-5c50-4e85-bb48-a099cfbb892d\") " Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.060682 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "710d1522-5c50-4e85-bb48-a099cfbb892d" (UID: "710d1522-5c50-4e85-bb48-a099cfbb892d"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.060967 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/710d1522-5c50-4e85-bb48-a099cfbb892d-kube-api-access-6cwfm" (OuterVolumeSpecName: "kube-api-access-6cwfm") pod "710d1522-5c50-4e85-bb48-a099cfbb892d" (UID: "710d1522-5c50-4e85-bb48-a099cfbb892d"). InnerVolumeSpecName "kube-api-access-6cwfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.081270 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "710d1522-5c50-4e85-bb48-a099cfbb892d" (UID: "710d1522-5c50-4e85-bb48-a099cfbb892d"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.093680 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-inventory" (OuterVolumeSpecName: "inventory") pod "710d1522-5c50-4e85-bb48-a099cfbb892d" (UID: "710d1522-5c50-4e85-bb48-a099cfbb892d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.099680 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "710d1522-5c50-4e85-bb48-a099cfbb892d" (UID: "710d1522-5c50-4e85-bb48-a099cfbb892d"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.105441 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "710d1522-5c50-4e85-bb48-a099cfbb892d" (UID: "710d1522-5c50-4e85-bb48-a099cfbb892d"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.118438 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "710d1522-5c50-4e85-bb48-a099cfbb892d" (UID: "710d1522-5c50-4e85-bb48-a099cfbb892d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.119023 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "710d1522-5c50-4e85-bb48-a099cfbb892d" (UID: "710d1522-5c50-4e85-bb48-a099cfbb892d"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.123660 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "710d1522-5c50-4e85-bb48-a099cfbb892d" (UID: "710d1522-5c50-4e85-bb48-a099cfbb892d"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.151826 4744 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.151863 4744 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.151872 4744 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.151882 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.151917 4744 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.151928 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.151937 4744 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.151945 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cwfm\" (UniqueName: \"kubernetes.io/projected/710d1522-5c50-4e85-bb48-a099cfbb892d-kube-api-access-6cwfm\") on node \"crc\" DevicePath \"\"" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.151954 4744 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/710d1522-5c50-4e85-bb48-a099cfbb892d-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.483587 4744 generic.go:334] "Generic (PLEG): container finished" podID="5209fede-d923-45ff-9f9e-b17962c687b2" containerID="1430bd17ca791adefac7c53cf7b38050548e5fb2a5a9b3daef2701edf2b83275" exitCode=0 Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.483642 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx9fw" event={"ID":"5209fede-d923-45ff-9f9e-b17962c687b2","Type":"ContainerDied","Data":"1430bd17ca791adefac7c53cf7b38050548e5fb2a5a9b3daef2701edf2b83275"} Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.486722 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" event={"ID":"710d1522-5c50-4e85-bb48-a099cfbb892d","Type":"ContainerDied","Data":"76d2040728269f6e9b638c488e4d7fcbfd678b4fbcbdcca16934641e4226da7f"} Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.486784 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76d2040728269f6e9b638c488e4d7fcbfd678b4fbcbdcca16934641e4226da7f" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.486740 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-z279b" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.666165 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn"] Oct 03 17:07:35 crc kubenswrapper[4744]: E1003 17:07:35.666612 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="710d1522-5c50-4e85-bb48-a099cfbb892d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.666635 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="710d1522-5c50-4e85-bb48-a099cfbb892d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.666884 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="710d1522-5c50-4e85-bb48-a099cfbb892d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.667850 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.671652 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.671882 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.685223 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn"] Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.706442 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-srfc9" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.706459 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.706624 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.765170 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs6qq\" (UniqueName: \"kubernetes.io/projected/4d4a0574-77fd-40fe-aa97-c4a26255473e-kube-api-access-qs6qq\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.765292 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.765331 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.765373 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.765403 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.765423 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.765695 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.868814 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qs6qq\" (UniqueName: \"kubernetes.io/projected/4d4a0574-77fd-40fe-aa97-c4a26255473e-kube-api-access-qs6qq\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.868952 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.869025 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.869089 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.869122 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.869142 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.869191 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.874071 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.874375 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.875102 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.884965 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.885067 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.888229 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:35 crc kubenswrapper[4744]: I1003 17:07:35.900372 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qs6qq\" (UniqueName: \"kubernetes.io/projected/4d4a0574-77fd-40fe-aa97-c4a26255473e-kube-api-access-qs6qq\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:36 crc kubenswrapper[4744]: I1003 17:07:36.020168 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:07:36 crc kubenswrapper[4744]: I1003 17:07:36.498619 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx9fw" event={"ID":"5209fede-d923-45ff-9f9e-b17962c687b2","Type":"ContainerStarted","Data":"915e1070da7d486691e9982407db0b9ee8d1cff1567a058654be607081a04644"} Oct 03 17:07:36 crc kubenswrapper[4744]: I1003 17:07:36.521860 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lx9fw" podStartSLOduration=3.072424277 podStartE2EDuration="5.521842249s" podCreationTimestamp="2025-10-03 17:07:31 +0000 UTC" firstStartedPulling="2025-10-03 17:07:33.464722422 +0000 UTC m=+2579.744598328" lastFinishedPulling="2025-10-03 17:07:35.914140394 +0000 UTC m=+2582.194016300" observedRunningTime="2025-10-03 17:07:36.520377304 +0000 UTC m=+2582.800253210" watchObservedRunningTime="2025-10-03 17:07:36.521842249 +0000 UTC m=+2582.801718155" Oct 03 17:07:36 crc kubenswrapper[4744]: W1003 17:07:36.570095 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d4a0574_77fd_40fe_aa97_c4a26255473e.slice/crio-cfaf9d1182b357b7d89b2432b621342ffcdea8ec9c75731d264045bc8cb1f7e3 WatchSource:0}: Error finding container cfaf9d1182b357b7d89b2432b621342ffcdea8ec9c75731d264045bc8cb1f7e3: Status 404 returned error can't find the container with id cfaf9d1182b357b7d89b2432b621342ffcdea8ec9c75731d264045bc8cb1f7e3 Oct 03 17:07:36 crc kubenswrapper[4744]: I1003 17:07:36.571666 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn"] Oct 03 17:07:37 crc kubenswrapper[4744]: I1003 17:07:37.512836 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" event={"ID":"4d4a0574-77fd-40fe-aa97-c4a26255473e","Type":"ContainerStarted","Data":"a9b5d60dbe6ce9c1ebf6f4ed9e2cc695c9a3ab53ebf40cda9f1b75e0f2932265"} Oct 03 17:07:37 crc kubenswrapper[4744]: I1003 17:07:37.513157 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" event={"ID":"4d4a0574-77fd-40fe-aa97-c4a26255473e","Type":"ContainerStarted","Data":"cfaf9d1182b357b7d89b2432b621342ffcdea8ec9c75731d264045bc8cb1f7e3"} Oct 03 17:07:37 crc kubenswrapper[4744]: I1003 17:07:37.534995 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" podStartSLOduration=2.04319997 podStartE2EDuration="2.534978602s" podCreationTimestamp="2025-10-03 17:07:35 +0000 UTC" firstStartedPulling="2025-10-03 17:07:36.572336454 +0000 UTC m=+2582.852212350" lastFinishedPulling="2025-10-03 17:07:37.064115066 +0000 UTC m=+2583.343990982" observedRunningTime="2025-10-03 17:07:37.528200471 +0000 UTC m=+2583.808076387" watchObservedRunningTime="2025-10-03 17:07:37.534978602 +0000 UTC m=+2583.814854498" Oct 03 17:07:40 crc kubenswrapper[4744]: I1003 17:07:40.893372 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:07:40 crc kubenswrapper[4744]: E1003 17:07:40.894452 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:07:41 crc kubenswrapper[4744]: I1003 17:07:41.924162 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:41 crc kubenswrapper[4744]: I1003 17:07:41.924729 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:41 crc kubenswrapper[4744]: I1003 17:07:41.997263 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:42 crc kubenswrapper[4744]: I1003 17:07:42.649815 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:42 crc kubenswrapper[4744]: I1003 17:07:42.724982 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lx9fw"] Oct 03 17:07:44 crc kubenswrapper[4744]: I1003 17:07:44.596653 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lx9fw" podUID="5209fede-d923-45ff-9f9e-b17962c687b2" containerName="registry-server" containerID="cri-o://915e1070da7d486691e9982407db0b9ee8d1cff1567a058654be607081a04644" gracePeriod=2 Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.121649 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.206628 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl8r5\" (UniqueName: \"kubernetes.io/projected/5209fede-d923-45ff-9f9e-b17962c687b2-kube-api-access-jl8r5\") pod \"5209fede-d923-45ff-9f9e-b17962c687b2\" (UID: \"5209fede-d923-45ff-9f9e-b17962c687b2\") " Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.206746 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5209fede-d923-45ff-9f9e-b17962c687b2-utilities\") pod \"5209fede-d923-45ff-9f9e-b17962c687b2\" (UID: \"5209fede-d923-45ff-9f9e-b17962c687b2\") " Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.206802 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5209fede-d923-45ff-9f9e-b17962c687b2-catalog-content\") pod \"5209fede-d923-45ff-9f9e-b17962c687b2\" (UID: \"5209fede-d923-45ff-9f9e-b17962c687b2\") " Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.208631 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5209fede-d923-45ff-9f9e-b17962c687b2-utilities" (OuterVolumeSpecName: "utilities") pod "5209fede-d923-45ff-9f9e-b17962c687b2" (UID: "5209fede-d923-45ff-9f9e-b17962c687b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.212810 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5209fede-d923-45ff-9f9e-b17962c687b2-kube-api-access-jl8r5" (OuterVolumeSpecName: "kube-api-access-jl8r5") pod "5209fede-d923-45ff-9f9e-b17962c687b2" (UID: "5209fede-d923-45ff-9f9e-b17962c687b2"). InnerVolumeSpecName "kube-api-access-jl8r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.308852 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl8r5\" (UniqueName: \"kubernetes.io/projected/5209fede-d923-45ff-9f9e-b17962c687b2-kube-api-access-jl8r5\") on node \"crc\" DevicePath \"\"" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.309082 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5209fede-d923-45ff-9f9e-b17962c687b2-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.509859 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5209fede-d923-45ff-9f9e-b17962c687b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5209fede-d923-45ff-9f9e-b17962c687b2" (UID: "5209fede-d923-45ff-9f9e-b17962c687b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.513609 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5209fede-d923-45ff-9f9e-b17962c687b2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.608199 4744 generic.go:334] "Generic (PLEG): container finished" podID="5209fede-d923-45ff-9f9e-b17962c687b2" containerID="915e1070da7d486691e9982407db0b9ee8d1cff1567a058654be607081a04644" exitCode=0 Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.608236 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx9fw" event={"ID":"5209fede-d923-45ff-9f9e-b17962c687b2","Type":"ContainerDied","Data":"915e1070da7d486691e9982407db0b9ee8d1cff1567a058654be607081a04644"} Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.608261 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx9fw" event={"ID":"5209fede-d923-45ff-9f9e-b17962c687b2","Type":"ContainerDied","Data":"1013dcc75d4e2a81a6c9b1ea74135f3547706c554d9ed6769752ac7cbbab4a4e"} Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.608277 4744 scope.go:117] "RemoveContainer" containerID="915e1070da7d486691e9982407db0b9ee8d1cff1567a058654be607081a04644" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.608397 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx9fw" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.644205 4744 scope.go:117] "RemoveContainer" containerID="1430bd17ca791adefac7c53cf7b38050548e5fb2a5a9b3daef2701edf2b83275" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.646279 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lx9fw"] Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.656977 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lx9fw"] Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.684360 4744 scope.go:117] "RemoveContainer" containerID="9aa4ff0e2b9dd21b80a626f3813971cb1d5810f18c44eb5f1b07f0a80bd5acfb" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.731088 4744 scope.go:117] "RemoveContainer" containerID="915e1070da7d486691e9982407db0b9ee8d1cff1567a058654be607081a04644" Oct 03 17:07:45 crc kubenswrapper[4744]: E1003 17:07:45.731539 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"915e1070da7d486691e9982407db0b9ee8d1cff1567a058654be607081a04644\": container with ID starting with 915e1070da7d486691e9982407db0b9ee8d1cff1567a058654be607081a04644 not found: ID does not exist" containerID="915e1070da7d486691e9982407db0b9ee8d1cff1567a058654be607081a04644" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.731568 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"915e1070da7d486691e9982407db0b9ee8d1cff1567a058654be607081a04644"} err="failed to get container status \"915e1070da7d486691e9982407db0b9ee8d1cff1567a058654be607081a04644\": rpc error: code = NotFound desc = could not find container \"915e1070da7d486691e9982407db0b9ee8d1cff1567a058654be607081a04644\": container with ID starting with 915e1070da7d486691e9982407db0b9ee8d1cff1567a058654be607081a04644 not found: ID does not exist" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.731590 4744 scope.go:117] "RemoveContainer" containerID="1430bd17ca791adefac7c53cf7b38050548e5fb2a5a9b3daef2701edf2b83275" Oct 03 17:07:45 crc kubenswrapper[4744]: E1003 17:07:45.732559 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1430bd17ca791adefac7c53cf7b38050548e5fb2a5a9b3daef2701edf2b83275\": container with ID starting with 1430bd17ca791adefac7c53cf7b38050548e5fb2a5a9b3daef2701edf2b83275 not found: ID does not exist" containerID="1430bd17ca791adefac7c53cf7b38050548e5fb2a5a9b3daef2701edf2b83275" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.732631 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1430bd17ca791adefac7c53cf7b38050548e5fb2a5a9b3daef2701edf2b83275"} err="failed to get container status \"1430bd17ca791adefac7c53cf7b38050548e5fb2a5a9b3daef2701edf2b83275\": rpc error: code = NotFound desc = could not find container \"1430bd17ca791adefac7c53cf7b38050548e5fb2a5a9b3daef2701edf2b83275\": container with ID starting with 1430bd17ca791adefac7c53cf7b38050548e5fb2a5a9b3daef2701edf2b83275 not found: ID does not exist" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.732660 4744 scope.go:117] "RemoveContainer" containerID="9aa4ff0e2b9dd21b80a626f3813971cb1d5810f18c44eb5f1b07f0a80bd5acfb" Oct 03 17:07:45 crc kubenswrapper[4744]: E1003 17:07:45.732950 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9aa4ff0e2b9dd21b80a626f3813971cb1d5810f18c44eb5f1b07f0a80bd5acfb\": container with ID starting with 9aa4ff0e2b9dd21b80a626f3813971cb1d5810f18c44eb5f1b07f0a80bd5acfb not found: ID does not exist" containerID="9aa4ff0e2b9dd21b80a626f3813971cb1d5810f18c44eb5f1b07f0a80bd5acfb" Oct 03 17:07:45 crc kubenswrapper[4744]: I1003 17:07:45.732977 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aa4ff0e2b9dd21b80a626f3813971cb1d5810f18c44eb5f1b07f0a80bd5acfb"} err="failed to get container status \"9aa4ff0e2b9dd21b80a626f3813971cb1d5810f18c44eb5f1b07f0a80bd5acfb\": rpc error: code = NotFound desc = could not find container \"9aa4ff0e2b9dd21b80a626f3813971cb1d5810f18c44eb5f1b07f0a80bd5acfb\": container with ID starting with 9aa4ff0e2b9dd21b80a626f3813971cb1d5810f18c44eb5f1b07f0a80bd5acfb not found: ID does not exist" Oct 03 17:07:46 crc kubenswrapper[4744]: I1003 17:07:46.911618 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5209fede-d923-45ff-9f9e-b17962c687b2" path="/var/lib/kubelet/pods/5209fede-d923-45ff-9f9e-b17962c687b2/volumes" Oct 03 17:07:52 crc kubenswrapper[4744]: I1003 17:07:52.892592 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:07:52 crc kubenswrapper[4744]: E1003 17:07:52.893471 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:08:03 crc kubenswrapper[4744]: I1003 17:08:03.892989 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:08:03 crc kubenswrapper[4744]: E1003 17:08:03.894450 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:08:15 crc kubenswrapper[4744]: I1003 17:08:15.892021 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:08:15 crc kubenswrapper[4744]: E1003 17:08:15.892669 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:08:16 crc kubenswrapper[4744]: I1003 17:08:16.995631 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lv5t5"] Oct 03 17:08:16 crc kubenswrapper[4744]: E1003 17:08:16.996619 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5209fede-d923-45ff-9f9e-b17962c687b2" containerName="extract-content" Oct 03 17:08:16 crc kubenswrapper[4744]: I1003 17:08:16.996640 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5209fede-d923-45ff-9f9e-b17962c687b2" containerName="extract-content" Oct 03 17:08:16 crc kubenswrapper[4744]: E1003 17:08:16.996673 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5209fede-d923-45ff-9f9e-b17962c687b2" containerName="registry-server" Oct 03 17:08:16 crc kubenswrapper[4744]: I1003 17:08:16.996683 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5209fede-d923-45ff-9f9e-b17962c687b2" containerName="registry-server" Oct 03 17:08:16 crc kubenswrapper[4744]: E1003 17:08:16.996717 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5209fede-d923-45ff-9f9e-b17962c687b2" containerName="extract-utilities" Oct 03 17:08:16 crc kubenswrapper[4744]: I1003 17:08:16.996729 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5209fede-d923-45ff-9f9e-b17962c687b2" containerName="extract-utilities" Oct 03 17:08:16 crc kubenswrapper[4744]: I1003 17:08:16.997082 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5209fede-d923-45ff-9f9e-b17962c687b2" containerName="registry-server" Oct 03 17:08:16 crc kubenswrapper[4744]: I1003 17:08:16.999384 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:17 crc kubenswrapper[4744]: I1003 17:08:17.013331 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lv5t5"] Oct 03 17:08:17 crc kubenswrapper[4744]: I1003 17:08:17.134971 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb394ad8-cdac-46ca-811a-04a56dec16d3-catalog-content\") pod \"community-operators-lv5t5\" (UID: \"cb394ad8-cdac-46ca-811a-04a56dec16d3\") " pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:17 crc kubenswrapper[4744]: I1003 17:08:17.135198 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb394ad8-cdac-46ca-811a-04a56dec16d3-utilities\") pod \"community-operators-lv5t5\" (UID: \"cb394ad8-cdac-46ca-811a-04a56dec16d3\") " pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:17 crc kubenswrapper[4744]: I1003 17:08:17.135333 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hxbk\" (UniqueName: \"kubernetes.io/projected/cb394ad8-cdac-46ca-811a-04a56dec16d3-kube-api-access-9hxbk\") pod \"community-operators-lv5t5\" (UID: \"cb394ad8-cdac-46ca-811a-04a56dec16d3\") " pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:17 crc kubenswrapper[4744]: I1003 17:08:17.237671 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb394ad8-cdac-46ca-811a-04a56dec16d3-catalog-content\") pod \"community-operators-lv5t5\" (UID: \"cb394ad8-cdac-46ca-811a-04a56dec16d3\") " pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:17 crc kubenswrapper[4744]: I1003 17:08:17.237796 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb394ad8-cdac-46ca-811a-04a56dec16d3-utilities\") pod \"community-operators-lv5t5\" (UID: \"cb394ad8-cdac-46ca-811a-04a56dec16d3\") " pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:17 crc kubenswrapper[4744]: I1003 17:08:17.237870 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hxbk\" (UniqueName: \"kubernetes.io/projected/cb394ad8-cdac-46ca-811a-04a56dec16d3-kube-api-access-9hxbk\") pod \"community-operators-lv5t5\" (UID: \"cb394ad8-cdac-46ca-811a-04a56dec16d3\") " pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:17 crc kubenswrapper[4744]: I1003 17:08:17.238350 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb394ad8-cdac-46ca-811a-04a56dec16d3-catalog-content\") pod \"community-operators-lv5t5\" (UID: \"cb394ad8-cdac-46ca-811a-04a56dec16d3\") " pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:17 crc kubenswrapper[4744]: I1003 17:08:17.238399 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb394ad8-cdac-46ca-811a-04a56dec16d3-utilities\") pod \"community-operators-lv5t5\" (UID: \"cb394ad8-cdac-46ca-811a-04a56dec16d3\") " pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:17 crc kubenswrapper[4744]: I1003 17:08:17.260973 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hxbk\" (UniqueName: \"kubernetes.io/projected/cb394ad8-cdac-46ca-811a-04a56dec16d3-kube-api-access-9hxbk\") pod \"community-operators-lv5t5\" (UID: \"cb394ad8-cdac-46ca-811a-04a56dec16d3\") " pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:17 crc kubenswrapper[4744]: I1003 17:08:17.323372 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:17 crc kubenswrapper[4744]: I1003 17:08:17.872241 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lv5t5"] Oct 03 17:08:17 crc kubenswrapper[4744]: I1003 17:08:17.975670 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv5t5" event={"ID":"cb394ad8-cdac-46ca-811a-04a56dec16d3","Type":"ContainerStarted","Data":"ca1534aec65a2c4e0a2259d1635afea50f876e72d252cf98cb88219bf03591f9"} Oct 03 17:08:18 crc kubenswrapper[4744]: I1003 17:08:18.988207 4744 generic.go:334] "Generic (PLEG): container finished" podID="cb394ad8-cdac-46ca-811a-04a56dec16d3" containerID="6b552c23e73bf1b9202e664f796b9a19704f0f3185067b59d1549359b2ff1813" exitCode=0 Oct 03 17:08:18 crc kubenswrapper[4744]: I1003 17:08:18.988578 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv5t5" event={"ID":"cb394ad8-cdac-46ca-811a-04a56dec16d3","Type":"ContainerDied","Data":"6b552c23e73bf1b9202e664f796b9a19704f0f3185067b59d1549359b2ff1813"} Oct 03 17:08:19 crc kubenswrapper[4744]: I1003 17:08:19.999112 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv5t5" event={"ID":"cb394ad8-cdac-46ca-811a-04a56dec16d3","Type":"ContainerStarted","Data":"5d815d5c6f38db34018af1fe7ba22e68abaacc6572f6bd8803ce71d386423f2f"} Oct 03 17:08:21 crc kubenswrapper[4744]: I1003 17:08:21.012332 4744 generic.go:334] "Generic (PLEG): container finished" podID="cb394ad8-cdac-46ca-811a-04a56dec16d3" containerID="5d815d5c6f38db34018af1fe7ba22e68abaacc6572f6bd8803ce71d386423f2f" exitCode=0 Oct 03 17:08:21 crc kubenswrapper[4744]: I1003 17:08:21.012450 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv5t5" event={"ID":"cb394ad8-cdac-46ca-811a-04a56dec16d3","Type":"ContainerDied","Data":"5d815d5c6f38db34018af1fe7ba22e68abaacc6572f6bd8803ce71d386423f2f"} Oct 03 17:08:22 crc kubenswrapper[4744]: I1003 17:08:22.045016 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv5t5" event={"ID":"cb394ad8-cdac-46ca-811a-04a56dec16d3","Type":"ContainerStarted","Data":"5323acc66dcb3d4a2bc27f952e9736586f5f93b82a8836f5cedf0907906b1859"} Oct 03 17:08:22 crc kubenswrapper[4744]: I1003 17:08:22.071168 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lv5t5" podStartSLOduration=3.486322511 podStartE2EDuration="6.071146579s" podCreationTimestamp="2025-10-03 17:08:16 +0000 UTC" firstStartedPulling="2025-10-03 17:08:18.991306383 +0000 UTC m=+2625.271182309" lastFinishedPulling="2025-10-03 17:08:21.576130451 +0000 UTC m=+2627.856006377" observedRunningTime="2025-10-03 17:08:22.069859749 +0000 UTC m=+2628.349735655" watchObservedRunningTime="2025-10-03 17:08:22.071146579 +0000 UTC m=+2628.351022475" Oct 03 17:08:27 crc kubenswrapper[4744]: I1003 17:08:27.324221 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:27 crc kubenswrapper[4744]: I1003 17:08:27.325444 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:27 crc kubenswrapper[4744]: I1003 17:08:27.396601 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:28 crc kubenswrapper[4744]: I1003 17:08:28.168390 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:28 crc kubenswrapper[4744]: I1003 17:08:28.224379 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lv5t5"] Oct 03 17:08:30 crc kubenswrapper[4744]: I1003 17:08:30.122767 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lv5t5" podUID="cb394ad8-cdac-46ca-811a-04a56dec16d3" containerName="registry-server" containerID="cri-o://5323acc66dcb3d4a2bc27f952e9736586f5f93b82a8836f5cedf0907906b1859" gracePeriod=2 Oct 03 17:08:30 crc kubenswrapper[4744]: I1003 17:08:30.627784 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:30 crc kubenswrapper[4744]: I1003 17:08:30.765786 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb394ad8-cdac-46ca-811a-04a56dec16d3-utilities\") pod \"cb394ad8-cdac-46ca-811a-04a56dec16d3\" (UID: \"cb394ad8-cdac-46ca-811a-04a56dec16d3\") " Oct 03 17:08:30 crc kubenswrapper[4744]: I1003 17:08:30.766026 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb394ad8-cdac-46ca-811a-04a56dec16d3-catalog-content\") pod \"cb394ad8-cdac-46ca-811a-04a56dec16d3\" (UID: \"cb394ad8-cdac-46ca-811a-04a56dec16d3\") " Oct 03 17:08:30 crc kubenswrapper[4744]: I1003 17:08:30.766053 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hxbk\" (UniqueName: \"kubernetes.io/projected/cb394ad8-cdac-46ca-811a-04a56dec16d3-kube-api-access-9hxbk\") pod \"cb394ad8-cdac-46ca-811a-04a56dec16d3\" (UID: \"cb394ad8-cdac-46ca-811a-04a56dec16d3\") " Oct 03 17:08:30 crc kubenswrapper[4744]: I1003 17:08:30.767171 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb394ad8-cdac-46ca-811a-04a56dec16d3-utilities" (OuterVolumeSpecName: "utilities") pod "cb394ad8-cdac-46ca-811a-04a56dec16d3" (UID: "cb394ad8-cdac-46ca-811a-04a56dec16d3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:08:30 crc kubenswrapper[4744]: I1003 17:08:30.776716 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb394ad8-cdac-46ca-811a-04a56dec16d3-kube-api-access-9hxbk" (OuterVolumeSpecName: "kube-api-access-9hxbk") pod "cb394ad8-cdac-46ca-811a-04a56dec16d3" (UID: "cb394ad8-cdac-46ca-811a-04a56dec16d3"). InnerVolumeSpecName "kube-api-access-9hxbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:08:30 crc kubenswrapper[4744]: I1003 17:08:30.868440 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hxbk\" (UniqueName: \"kubernetes.io/projected/cb394ad8-cdac-46ca-811a-04a56dec16d3-kube-api-access-9hxbk\") on node \"crc\" DevicePath \"\"" Oct 03 17:08:30 crc kubenswrapper[4744]: I1003 17:08:30.868520 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb394ad8-cdac-46ca-811a-04a56dec16d3-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:08:30 crc kubenswrapper[4744]: I1003 17:08:30.893227 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:08:30 crc kubenswrapper[4744]: E1003 17:08:30.893880 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.017439 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb394ad8-cdac-46ca-811a-04a56dec16d3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb394ad8-cdac-46ca-811a-04a56dec16d3" (UID: "cb394ad8-cdac-46ca-811a-04a56dec16d3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.074398 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb394ad8-cdac-46ca-811a-04a56dec16d3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.155776 4744 generic.go:334] "Generic (PLEG): container finished" podID="cb394ad8-cdac-46ca-811a-04a56dec16d3" containerID="5323acc66dcb3d4a2bc27f952e9736586f5f93b82a8836f5cedf0907906b1859" exitCode=0 Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.155815 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv5t5" event={"ID":"cb394ad8-cdac-46ca-811a-04a56dec16d3","Type":"ContainerDied","Data":"5323acc66dcb3d4a2bc27f952e9736586f5f93b82a8836f5cedf0907906b1859"} Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.155840 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lv5t5" event={"ID":"cb394ad8-cdac-46ca-811a-04a56dec16d3","Type":"ContainerDied","Data":"ca1534aec65a2c4e0a2259d1635afea50f876e72d252cf98cb88219bf03591f9"} Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.155869 4744 scope.go:117] "RemoveContainer" containerID="5323acc66dcb3d4a2bc27f952e9736586f5f93b82a8836f5cedf0907906b1859" Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.155982 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lv5t5" Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.199582 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lv5t5"] Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.208919 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lv5t5"] Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.216227 4744 scope.go:117] "RemoveContainer" containerID="5d815d5c6f38db34018af1fe7ba22e68abaacc6572f6bd8803ce71d386423f2f" Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.239699 4744 scope.go:117] "RemoveContainer" containerID="6b552c23e73bf1b9202e664f796b9a19704f0f3185067b59d1549359b2ff1813" Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.295184 4744 scope.go:117] "RemoveContainer" containerID="5323acc66dcb3d4a2bc27f952e9736586f5f93b82a8836f5cedf0907906b1859" Oct 03 17:08:31 crc kubenswrapper[4744]: E1003 17:08:31.295749 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5323acc66dcb3d4a2bc27f952e9736586f5f93b82a8836f5cedf0907906b1859\": container with ID starting with 5323acc66dcb3d4a2bc27f952e9736586f5f93b82a8836f5cedf0907906b1859 not found: ID does not exist" containerID="5323acc66dcb3d4a2bc27f952e9736586f5f93b82a8836f5cedf0907906b1859" Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.295787 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5323acc66dcb3d4a2bc27f952e9736586f5f93b82a8836f5cedf0907906b1859"} err="failed to get container status \"5323acc66dcb3d4a2bc27f952e9736586f5f93b82a8836f5cedf0907906b1859\": rpc error: code = NotFound desc = could not find container \"5323acc66dcb3d4a2bc27f952e9736586f5f93b82a8836f5cedf0907906b1859\": container with ID starting with 5323acc66dcb3d4a2bc27f952e9736586f5f93b82a8836f5cedf0907906b1859 not found: ID does not exist" Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.295818 4744 scope.go:117] "RemoveContainer" containerID="5d815d5c6f38db34018af1fe7ba22e68abaacc6572f6bd8803ce71d386423f2f" Oct 03 17:08:31 crc kubenswrapper[4744]: E1003 17:08:31.296167 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d815d5c6f38db34018af1fe7ba22e68abaacc6572f6bd8803ce71d386423f2f\": container with ID starting with 5d815d5c6f38db34018af1fe7ba22e68abaacc6572f6bd8803ce71d386423f2f not found: ID does not exist" containerID="5d815d5c6f38db34018af1fe7ba22e68abaacc6572f6bd8803ce71d386423f2f" Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.296193 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d815d5c6f38db34018af1fe7ba22e68abaacc6572f6bd8803ce71d386423f2f"} err="failed to get container status \"5d815d5c6f38db34018af1fe7ba22e68abaacc6572f6bd8803ce71d386423f2f\": rpc error: code = NotFound desc = could not find container \"5d815d5c6f38db34018af1fe7ba22e68abaacc6572f6bd8803ce71d386423f2f\": container with ID starting with 5d815d5c6f38db34018af1fe7ba22e68abaacc6572f6bd8803ce71d386423f2f not found: ID does not exist" Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.296218 4744 scope.go:117] "RemoveContainer" containerID="6b552c23e73bf1b9202e664f796b9a19704f0f3185067b59d1549359b2ff1813" Oct 03 17:08:31 crc kubenswrapper[4744]: E1003 17:08:31.296798 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b552c23e73bf1b9202e664f796b9a19704f0f3185067b59d1549359b2ff1813\": container with ID starting with 6b552c23e73bf1b9202e664f796b9a19704f0f3185067b59d1549359b2ff1813 not found: ID does not exist" containerID="6b552c23e73bf1b9202e664f796b9a19704f0f3185067b59d1549359b2ff1813" Oct 03 17:08:31 crc kubenswrapper[4744]: I1003 17:08:31.296834 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b552c23e73bf1b9202e664f796b9a19704f0f3185067b59d1549359b2ff1813"} err="failed to get container status \"6b552c23e73bf1b9202e664f796b9a19704f0f3185067b59d1549359b2ff1813\": rpc error: code = NotFound desc = could not find container \"6b552c23e73bf1b9202e664f796b9a19704f0f3185067b59d1549359b2ff1813\": container with ID starting with 6b552c23e73bf1b9202e664f796b9a19704f0f3185067b59d1549359b2ff1813 not found: ID does not exist" Oct 03 17:08:32 crc kubenswrapper[4744]: I1003 17:08:32.906222 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb394ad8-cdac-46ca-811a-04a56dec16d3" path="/var/lib/kubelet/pods/cb394ad8-cdac-46ca-811a-04a56dec16d3/volumes" Oct 03 17:08:44 crc kubenswrapper[4744]: I1003 17:08:44.908116 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:08:45 crc kubenswrapper[4744]: I1003 17:08:45.336778 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"5e65036ce88f44b88c0f29775217f2811e8ec53114ae1711b6cfd674a3ffb8ab"} Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.159178 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vkm2q"] Oct 03 17:09:52 crc kubenswrapper[4744]: E1003 17:09:52.160290 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb394ad8-cdac-46ca-811a-04a56dec16d3" containerName="extract-content" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.160305 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb394ad8-cdac-46ca-811a-04a56dec16d3" containerName="extract-content" Oct 03 17:09:52 crc kubenswrapper[4744]: E1003 17:09:52.160319 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb394ad8-cdac-46ca-811a-04a56dec16d3" containerName="registry-server" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.160327 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb394ad8-cdac-46ca-811a-04a56dec16d3" containerName="registry-server" Oct 03 17:09:52 crc kubenswrapper[4744]: E1003 17:09:52.160359 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb394ad8-cdac-46ca-811a-04a56dec16d3" containerName="extract-utilities" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.160367 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb394ad8-cdac-46ca-811a-04a56dec16d3" containerName="extract-utilities" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.160622 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb394ad8-cdac-46ca-811a-04a56dec16d3" containerName="registry-server" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.162216 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.178510 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vkm2q"] Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.309905 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-utilities\") pod \"redhat-operators-vkm2q\" (UID: \"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8\") " pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.310033 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-catalog-content\") pod \"redhat-operators-vkm2q\" (UID: \"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8\") " pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.310084 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nltgz\" (UniqueName: \"kubernetes.io/projected/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-kube-api-access-nltgz\") pod \"redhat-operators-vkm2q\" (UID: \"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8\") " pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.411944 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-utilities\") pod \"redhat-operators-vkm2q\" (UID: \"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8\") " pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.412248 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-catalog-content\") pod \"redhat-operators-vkm2q\" (UID: \"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8\") " pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.412294 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nltgz\" (UniqueName: \"kubernetes.io/projected/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-kube-api-access-nltgz\") pod \"redhat-operators-vkm2q\" (UID: \"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8\") " pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.412985 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-utilities\") pod \"redhat-operators-vkm2q\" (UID: \"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8\") " pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.413206 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-catalog-content\") pod \"redhat-operators-vkm2q\" (UID: \"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8\") " pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.433635 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nltgz\" (UniqueName: \"kubernetes.io/projected/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-kube-api-access-nltgz\") pod \"redhat-operators-vkm2q\" (UID: \"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8\") " pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.487782 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:09:52 crc kubenswrapper[4744]: I1003 17:09:52.994921 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vkm2q"] Oct 03 17:09:53 crc kubenswrapper[4744]: I1003 17:09:53.108394 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkm2q" event={"ID":"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8","Type":"ContainerStarted","Data":"6a0419892661ab8a0048386b6a39e56dae20414544146e53b1e98f40821bc32a"} Oct 03 17:09:54 crc kubenswrapper[4744]: I1003 17:09:54.125256 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkm2q" event={"ID":"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8","Type":"ContainerDied","Data":"72829211a1b056f1c13e431fb511fcb7e177952b87ee88c3a5f1fa38e4e9634f"} Oct 03 17:09:54 crc kubenswrapper[4744]: I1003 17:09:54.125191 4744 generic.go:334] "Generic (PLEG): container finished" podID="5b3d487a-c3c3-4032-81ee-e8c6b39a39a8" containerID="72829211a1b056f1c13e431fb511fcb7e177952b87ee88c3a5f1fa38e4e9634f" exitCode=0 Oct 03 17:09:54 crc kubenswrapper[4744]: I1003 17:09:54.127875 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 17:09:54 crc kubenswrapper[4744]: I1003 17:09:54.132210 4744 generic.go:334] "Generic (PLEG): container finished" podID="4d4a0574-77fd-40fe-aa97-c4a26255473e" containerID="a9b5d60dbe6ce9c1ebf6f4ed9e2cc695c9a3ab53ebf40cda9f1b75e0f2932265" exitCode=0 Oct 03 17:09:54 crc kubenswrapper[4744]: I1003 17:09:54.132256 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" event={"ID":"4d4a0574-77fd-40fe-aa97-c4a26255473e","Type":"ContainerDied","Data":"a9b5d60dbe6ce9c1ebf6f4ed9e2cc695c9a3ab53ebf40cda9f1b75e0f2932265"} Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.575859 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.676060 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs6qq\" (UniqueName: \"kubernetes.io/projected/4d4a0574-77fd-40fe-aa97-c4a26255473e-kube-api-access-qs6qq\") pod \"4d4a0574-77fd-40fe-aa97-c4a26255473e\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.676114 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-inventory\") pod \"4d4a0574-77fd-40fe-aa97-c4a26255473e\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.676172 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-2\") pod \"4d4a0574-77fd-40fe-aa97-c4a26255473e\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.676273 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ssh-key\") pod \"4d4a0574-77fd-40fe-aa97-c4a26255473e\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.676316 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-telemetry-combined-ca-bundle\") pod \"4d4a0574-77fd-40fe-aa97-c4a26255473e\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.676420 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-0\") pod \"4d4a0574-77fd-40fe-aa97-c4a26255473e\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.676445 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-1\") pod \"4d4a0574-77fd-40fe-aa97-c4a26255473e\" (UID: \"4d4a0574-77fd-40fe-aa97-c4a26255473e\") " Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.682830 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d4a0574-77fd-40fe-aa97-c4a26255473e-kube-api-access-qs6qq" (OuterVolumeSpecName: "kube-api-access-qs6qq") pod "4d4a0574-77fd-40fe-aa97-c4a26255473e" (UID: "4d4a0574-77fd-40fe-aa97-c4a26255473e"). InnerVolumeSpecName "kube-api-access-qs6qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.688130 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "4d4a0574-77fd-40fe-aa97-c4a26255473e" (UID: "4d4a0574-77fd-40fe-aa97-c4a26255473e"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.712273 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4d4a0574-77fd-40fe-aa97-c4a26255473e" (UID: "4d4a0574-77fd-40fe-aa97-c4a26255473e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.713007 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "4d4a0574-77fd-40fe-aa97-c4a26255473e" (UID: "4d4a0574-77fd-40fe-aa97-c4a26255473e"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.713808 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "4d4a0574-77fd-40fe-aa97-c4a26255473e" (UID: "4d4a0574-77fd-40fe-aa97-c4a26255473e"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.714301 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-inventory" (OuterVolumeSpecName: "inventory") pod "4d4a0574-77fd-40fe-aa97-c4a26255473e" (UID: "4d4a0574-77fd-40fe-aa97-c4a26255473e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.736019 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "4d4a0574-77fd-40fe-aa97-c4a26255473e" (UID: "4d4a0574-77fd-40fe-aa97-c4a26255473e"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.780157 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.780394 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.780532 4744 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.780642 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.780783 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.780914 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs6qq\" (UniqueName: \"kubernetes.io/projected/4d4a0574-77fd-40fe-aa97-c4a26255473e-kube-api-access-qs6qq\") on node \"crc\" DevicePath \"\"" Oct 03 17:09:55 crc kubenswrapper[4744]: I1003 17:09:55.781043 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d4a0574-77fd-40fe-aa97-c4a26255473e-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 17:09:56 crc kubenswrapper[4744]: I1003 17:09:56.149580 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" event={"ID":"4d4a0574-77fd-40fe-aa97-c4a26255473e","Type":"ContainerDied","Data":"cfaf9d1182b357b7d89b2432b621342ffcdea8ec9c75731d264045bc8cb1f7e3"} Oct 03 17:09:56 crc kubenswrapper[4744]: I1003 17:09:56.149780 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfaf9d1182b357b7d89b2432b621342ffcdea8ec9c75731d264045bc8cb1f7e3" Oct 03 17:09:56 crc kubenswrapper[4744]: I1003 17:09:56.149669 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn" Oct 03 17:09:56 crc kubenswrapper[4744]: I1003 17:09:56.151732 4744 generic.go:334] "Generic (PLEG): container finished" podID="5b3d487a-c3c3-4032-81ee-e8c6b39a39a8" containerID="4e346de4acd7a750f4199a739351c124d3468857ff8643c8f38bf9841df10636" exitCode=0 Oct 03 17:09:56 crc kubenswrapper[4744]: I1003 17:09:56.151767 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkm2q" event={"ID":"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8","Type":"ContainerDied","Data":"4e346de4acd7a750f4199a739351c124d3468857ff8643c8f38bf9841df10636"} Oct 03 17:09:57 crc kubenswrapper[4744]: I1003 17:09:57.163300 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkm2q" event={"ID":"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8","Type":"ContainerStarted","Data":"f4fec8243573c97c58609c87fd7baae1f66c7cc35489182cb38887cff7053eed"} Oct 03 17:09:57 crc kubenswrapper[4744]: I1003 17:09:57.190420 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vkm2q" podStartSLOduration=2.63845124 podStartE2EDuration="5.190403767s" podCreationTimestamp="2025-10-03 17:09:52 +0000 UTC" firstStartedPulling="2025-10-03 17:09:54.127097775 +0000 UTC m=+2720.406973711" lastFinishedPulling="2025-10-03 17:09:56.679050302 +0000 UTC m=+2722.958926238" observedRunningTime="2025-10-03 17:09:57.186675624 +0000 UTC m=+2723.466551520" watchObservedRunningTime="2025-10-03 17:09:57.190403767 +0000 UTC m=+2723.470279653" Oct 03 17:10:02 crc kubenswrapper[4744]: I1003 17:10:02.488348 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:10:02 crc kubenswrapper[4744]: I1003 17:10:02.489761 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:10:02 crc kubenswrapper[4744]: I1003 17:10:02.557855 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:10:03 crc kubenswrapper[4744]: I1003 17:10:03.277541 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:10:03 crc kubenswrapper[4744]: I1003 17:10:03.330658 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vkm2q"] Oct 03 17:10:05 crc kubenswrapper[4744]: I1003 17:10:05.236950 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vkm2q" podUID="5b3d487a-c3c3-4032-81ee-e8c6b39a39a8" containerName="registry-server" containerID="cri-o://f4fec8243573c97c58609c87fd7baae1f66c7cc35489182cb38887cff7053eed" gracePeriod=2 Oct 03 17:10:05 crc kubenswrapper[4744]: I1003 17:10:05.659171 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:10:05 crc kubenswrapper[4744]: I1003 17:10:05.800406 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-catalog-content\") pod \"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8\" (UID: \"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8\") " Oct 03 17:10:05 crc kubenswrapper[4744]: I1003 17:10:05.801254 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-utilities\") pod \"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8\" (UID: \"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8\") " Oct 03 17:10:05 crc kubenswrapper[4744]: I1003 17:10:05.801449 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nltgz\" (UniqueName: \"kubernetes.io/projected/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-kube-api-access-nltgz\") pod \"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8\" (UID: \"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8\") " Oct 03 17:10:05 crc kubenswrapper[4744]: I1003 17:10:05.802244 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-utilities" (OuterVolumeSpecName: "utilities") pod "5b3d487a-c3c3-4032-81ee-e8c6b39a39a8" (UID: "5b3d487a-c3c3-4032-81ee-e8c6b39a39a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:10:05 crc kubenswrapper[4744]: I1003 17:10:05.803611 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:10:05 crc kubenswrapper[4744]: I1003 17:10:05.812355 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-kube-api-access-nltgz" (OuterVolumeSpecName: "kube-api-access-nltgz") pod "5b3d487a-c3c3-4032-81ee-e8c6b39a39a8" (UID: "5b3d487a-c3c3-4032-81ee-e8c6b39a39a8"). InnerVolumeSpecName "kube-api-access-nltgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:10:05 crc kubenswrapper[4744]: I1003 17:10:05.884152 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b3d487a-c3c3-4032-81ee-e8c6b39a39a8" (UID: "5b3d487a-c3c3-4032-81ee-e8c6b39a39a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:10:05 crc kubenswrapper[4744]: I1003 17:10:05.906029 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nltgz\" (UniqueName: \"kubernetes.io/projected/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-kube-api-access-nltgz\") on node \"crc\" DevicePath \"\"" Oct 03 17:10:05 crc kubenswrapper[4744]: I1003 17:10:05.906091 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.251478 4744 generic.go:334] "Generic (PLEG): container finished" podID="5b3d487a-c3c3-4032-81ee-e8c6b39a39a8" containerID="f4fec8243573c97c58609c87fd7baae1f66c7cc35489182cb38887cff7053eed" exitCode=0 Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.251559 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkm2q" event={"ID":"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8","Type":"ContainerDied","Data":"f4fec8243573c97c58609c87fd7baae1f66c7cc35489182cb38887cff7053eed"} Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.251597 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vkm2q" event={"ID":"5b3d487a-c3c3-4032-81ee-e8c6b39a39a8","Type":"ContainerDied","Data":"6a0419892661ab8a0048386b6a39e56dae20414544146e53b1e98f40821bc32a"} Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.251626 4744 scope.go:117] "RemoveContainer" containerID="f4fec8243573c97c58609c87fd7baae1f66c7cc35489182cb38887cff7053eed" Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.251822 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vkm2q" Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.284881 4744 scope.go:117] "RemoveContainer" containerID="4e346de4acd7a750f4199a739351c124d3468857ff8643c8f38bf9841df10636" Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.303603 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vkm2q"] Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.320357 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vkm2q"] Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.329391 4744 scope.go:117] "RemoveContainer" containerID="72829211a1b056f1c13e431fb511fcb7e177952b87ee88c3a5f1fa38e4e9634f" Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.386439 4744 scope.go:117] "RemoveContainer" containerID="f4fec8243573c97c58609c87fd7baae1f66c7cc35489182cb38887cff7053eed" Oct 03 17:10:06 crc kubenswrapper[4744]: E1003 17:10:06.386996 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4fec8243573c97c58609c87fd7baae1f66c7cc35489182cb38887cff7053eed\": container with ID starting with f4fec8243573c97c58609c87fd7baae1f66c7cc35489182cb38887cff7053eed not found: ID does not exist" containerID="f4fec8243573c97c58609c87fd7baae1f66c7cc35489182cb38887cff7053eed" Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.387037 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4fec8243573c97c58609c87fd7baae1f66c7cc35489182cb38887cff7053eed"} err="failed to get container status \"f4fec8243573c97c58609c87fd7baae1f66c7cc35489182cb38887cff7053eed\": rpc error: code = NotFound desc = could not find container \"f4fec8243573c97c58609c87fd7baae1f66c7cc35489182cb38887cff7053eed\": container with ID starting with f4fec8243573c97c58609c87fd7baae1f66c7cc35489182cb38887cff7053eed not found: ID does not exist" Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.387062 4744 scope.go:117] "RemoveContainer" containerID="4e346de4acd7a750f4199a739351c124d3468857ff8643c8f38bf9841df10636" Oct 03 17:10:06 crc kubenswrapper[4744]: E1003 17:10:06.387635 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e346de4acd7a750f4199a739351c124d3468857ff8643c8f38bf9841df10636\": container with ID starting with 4e346de4acd7a750f4199a739351c124d3468857ff8643c8f38bf9841df10636 not found: ID does not exist" containerID="4e346de4acd7a750f4199a739351c124d3468857ff8643c8f38bf9841df10636" Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.387674 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e346de4acd7a750f4199a739351c124d3468857ff8643c8f38bf9841df10636"} err="failed to get container status \"4e346de4acd7a750f4199a739351c124d3468857ff8643c8f38bf9841df10636\": rpc error: code = NotFound desc = could not find container \"4e346de4acd7a750f4199a739351c124d3468857ff8643c8f38bf9841df10636\": container with ID starting with 4e346de4acd7a750f4199a739351c124d3468857ff8643c8f38bf9841df10636 not found: ID does not exist" Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.387699 4744 scope.go:117] "RemoveContainer" containerID="72829211a1b056f1c13e431fb511fcb7e177952b87ee88c3a5f1fa38e4e9634f" Oct 03 17:10:06 crc kubenswrapper[4744]: E1003 17:10:06.388174 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72829211a1b056f1c13e431fb511fcb7e177952b87ee88c3a5f1fa38e4e9634f\": container with ID starting with 72829211a1b056f1c13e431fb511fcb7e177952b87ee88c3a5f1fa38e4e9634f not found: ID does not exist" containerID="72829211a1b056f1c13e431fb511fcb7e177952b87ee88c3a5f1fa38e4e9634f" Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.388206 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72829211a1b056f1c13e431fb511fcb7e177952b87ee88c3a5f1fa38e4e9634f"} err="failed to get container status \"72829211a1b056f1c13e431fb511fcb7e177952b87ee88c3a5f1fa38e4e9634f\": rpc error: code = NotFound desc = could not find container \"72829211a1b056f1c13e431fb511fcb7e177952b87ee88c3a5f1fa38e4e9634f\": container with ID starting with 72829211a1b056f1c13e431fb511fcb7e177952b87ee88c3a5f1fa38e4e9634f not found: ID does not exist" Oct 03 17:10:06 crc kubenswrapper[4744]: I1003 17:10:06.907792 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b3d487a-c3c3-4032-81ee-e8c6b39a39a8" path="/var/lib/kubelet/pods/5b3d487a-c3c3-4032-81ee-e8c6b39a39a8/volumes" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.407647 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 03 17:11:04 crc kubenswrapper[4744]: E1003 17:11:04.412043 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d4a0574-77fd-40fe-aa97-c4a26255473e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.412091 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d4a0574-77fd-40fe-aa97-c4a26255473e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 03 17:11:04 crc kubenswrapper[4744]: E1003 17:11:04.412123 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b3d487a-c3c3-4032-81ee-e8c6b39a39a8" containerName="extract-content" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.412151 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b3d487a-c3c3-4032-81ee-e8c6b39a39a8" containerName="extract-content" Oct 03 17:11:04 crc kubenswrapper[4744]: E1003 17:11:04.412193 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b3d487a-c3c3-4032-81ee-e8c6b39a39a8" containerName="extract-utilities" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.412204 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b3d487a-c3c3-4032-81ee-e8c6b39a39a8" containerName="extract-utilities" Oct 03 17:11:04 crc kubenswrapper[4744]: E1003 17:11:04.412220 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b3d487a-c3c3-4032-81ee-e8c6b39a39a8" containerName="registry-server" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.412231 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b3d487a-c3c3-4032-81ee-e8c6b39a39a8" containerName="registry-server" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.412608 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b3d487a-c3c3-4032-81ee-e8c6b39a39a8" containerName="registry-server" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.412658 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d4a0574-77fd-40fe-aa97-c4a26255473e" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.413682 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.418846 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.419246 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.419472 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.423790 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.494293 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/595f783e-db00-49b2-968f-79163001ffdd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.494366 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.494392 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.494439 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/595f783e-db00-49b2-968f-79163001ffdd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.494685 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.494790 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/595f783e-db00-49b2-968f-79163001ffdd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.494836 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b797f\" (UniqueName: \"kubernetes.io/projected/595f783e-db00-49b2-968f-79163001ffdd-kube-api-access-b797f\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.495122 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/595f783e-db00-49b2-968f-79163001ffdd-config-data\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.495234 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.597395 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/595f783e-db00-49b2-968f-79163001ffdd-config-data\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.597471 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.597553 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/595f783e-db00-49b2-968f-79163001ffdd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.597606 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.597636 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.597697 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/595f783e-db00-49b2-968f-79163001ffdd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.597762 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.597829 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/595f783e-db00-49b2-968f-79163001ffdd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.597868 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b797f\" (UniqueName: \"kubernetes.io/projected/595f783e-db00-49b2-968f-79163001ffdd-kube-api-access-b797f\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.598823 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/595f783e-db00-49b2-968f-79163001ffdd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.599136 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.599324 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/595f783e-db00-49b2-968f-79163001ffdd-config-data\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.599257 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/595f783e-db00-49b2-968f-79163001ffdd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.600807 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/595f783e-db00-49b2-968f-79163001ffdd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.606207 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.607032 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.614666 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.623396 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b797f\" (UniqueName: \"kubernetes.io/projected/595f783e-db00-49b2-968f-79163001ffdd-kube-api-access-b797f\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.642277 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " pod="openstack/tempest-tests-tempest" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.668341 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.668389 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:11:04 crc kubenswrapper[4744]: I1003 17:11:04.749143 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 03 17:11:05 crc kubenswrapper[4744]: I1003 17:11:05.034142 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 03 17:11:05 crc kubenswrapper[4744]: W1003 17:11:05.041268 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod595f783e_db00_49b2_968f_79163001ffdd.slice/crio-ce59043602b7ff62ea8d1aa8c885a946755ab4381b792d8f8b9ba6e442ed3b2f WatchSource:0}: Error finding container ce59043602b7ff62ea8d1aa8c885a946755ab4381b792d8f8b9ba6e442ed3b2f: Status 404 returned error can't find the container with id ce59043602b7ff62ea8d1aa8c885a946755ab4381b792d8f8b9ba6e442ed3b2f Oct 03 17:11:05 crc kubenswrapper[4744]: I1003 17:11:05.908305 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"595f783e-db00-49b2-968f-79163001ffdd","Type":"ContainerStarted","Data":"ce59043602b7ff62ea8d1aa8c885a946755ab4381b792d8f8b9ba6e442ed3b2f"} Oct 03 17:11:34 crc kubenswrapper[4744]: E1003 17:11:34.321459 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 03 17:11:34 crc kubenswrapper[4744]: E1003 17:11:34.322216 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b797f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(595f783e-db00-49b2-968f-79163001ffdd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 17:11:34 crc kubenswrapper[4744]: E1003 17:11:34.323440 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="595f783e-db00-49b2-968f-79163001ffdd" Oct 03 17:11:34 crc kubenswrapper[4744]: I1003 17:11:34.668885 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:11:34 crc kubenswrapper[4744]: I1003 17:11:34.668933 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:11:35 crc kubenswrapper[4744]: E1003 17:11:35.232585 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="595f783e-db00-49b2-968f-79163001ffdd" Oct 03 17:11:46 crc kubenswrapper[4744]: I1003 17:11:46.957010 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qmxd6"] Oct 03 17:11:46 crc kubenswrapper[4744]: I1003 17:11:46.964041 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:11:46 crc kubenswrapper[4744]: I1003 17:11:46.973182 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qmxd6"] Oct 03 17:11:47 crc kubenswrapper[4744]: I1003 17:11:47.096224 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9gwf\" (UniqueName: \"kubernetes.io/projected/68ad492e-af0b-4c93-9666-4f76715e8c2d-kube-api-access-n9gwf\") pod \"redhat-marketplace-qmxd6\" (UID: \"68ad492e-af0b-4c93-9666-4f76715e8c2d\") " pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:11:47 crc kubenswrapper[4744]: I1003 17:11:47.096667 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ad492e-af0b-4c93-9666-4f76715e8c2d-utilities\") pod \"redhat-marketplace-qmxd6\" (UID: \"68ad492e-af0b-4c93-9666-4f76715e8c2d\") " pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:11:47 crc kubenswrapper[4744]: I1003 17:11:47.096999 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ad492e-af0b-4c93-9666-4f76715e8c2d-catalog-content\") pod \"redhat-marketplace-qmxd6\" (UID: \"68ad492e-af0b-4c93-9666-4f76715e8c2d\") " pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:11:47 crc kubenswrapper[4744]: I1003 17:11:47.198945 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ad492e-af0b-4c93-9666-4f76715e8c2d-catalog-content\") pod \"redhat-marketplace-qmxd6\" (UID: \"68ad492e-af0b-4c93-9666-4f76715e8c2d\") " pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:11:47 crc kubenswrapper[4744]: I1003 17:11:47.199078 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9gwf\" (UniqueName: \"kubernetes.io/projected/68ad492e-af0b-4c93-9666-4f76715e8c2d-kube-api-access-n9gwf\") pod \"redhat-marketplace-qmxd6\" (UID: \"68ad492e-af0b-4c93-9666-4f76715e8c2d\") " pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:11:47 crc kubenswrapper[4744]: I1003 17:11:47.199213 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ad492e-af0b-4c93-9666-4f76715e8c2d-utilities\") pod \"redhat-marketplace-qmxd6\" (UID: \"68ad492e-af0b-4c93-9666-4f76715e8c2d\") " pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:11:47 crc kubenswrapper[4744]: I1003 17:11:47.199907 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ad492e-af0b-4c93-9666-4f76715e8c2d-utilities\") pod \"redhat-marketplace-qmxd6\" (UID: \"68ad492e-af0b-4c93-9666-4f76715e8c2d\") " pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:11:47 crc kubenswrapper[4744]: I1003 17:11:47.200072 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ad492e-af0b-4c93-9666-4f76715e8c2d-catalog-content\") pod \"redhat-marketplace-qmxd6\" (UID: \"68ad492e-af0b-4c93-9666-4f76715e8c2d\") " pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:11:47 crc kubenswrapper[4744]: I1003 17:11:47.234565 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9gwf\" (UniqueName: \"kubernetes.io/projected/68ad492e-af0b-4c93-9666-4f76715e8c2d-kube-api-access-n9gwf\") pod \"redhat-marketplace-qmxd6\" (UID: \"68ad492e-af0b-4c93-9666-4f76715e8c2d\") " pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:11:47 crc kubenswrapper[4744]: I1003 17:11:47.307786 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:11:47 crc kubenswrapper[4744]: I1003 17:11:47.801894 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qmxd6"] Oct 03 17:11:48 crc kubenswrapper[4744]: I1003 17:11:48.420413 4744 generic.go:334] "Generic (PLEG): container finished" podID="68ad492e-af0b-4c93-9666-4f76715e8c2d" containerID="3f1b8eecb5e4d879e58f295c2a35e8592d96ca33b3dc540d1fe2a6cbd2b0d705" exitCode=0 Oct 03 17:11:48 crc kubenswrapper[4744]: I1003 17:11:48.420542 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmxd6" event={"ID":"68ad492e-af0b-4c93-9666-4f76715e8c2d","Type":"ContainerDied","Data":"3f1b8eecb5e4d879e58f295c2a35e8592d96ca33b3dc540d1fe2a6cbd2b0d705"} Oct 03 17:11:48 crc kubenswrapper[4744]: I1003 17:11:48.421082 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmxd6" event={"ID":"68ad492e-af0b-4c93-9666-4f76715e8c2d","Type":"ContainerStarted","Data":"45568794ede4b8dcd6bcec127937814d01b12421a12361945d19ea260038bf6f"} Oct 03 17:11:50 crc kubenswrapper[4744]: I1003 17:11:50.479416 4744 generic.go:334] "Generic (PLEG): container finished" podID="68ad492e-af0b-4c93-9666-4f76715e8c2d" containerID="4634274d0596f237a6ff625c2528076f0e291464e7910506488e5f6db53eb461" exitCode=0 Oct 03 17:11:50 crc kubenswrapper[4744]: I1003 17:11:50.479624 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmxd6" event={"ID":"68ad492e-af0b-4c93-9666-4f76715e8c2d","Type":"ContainerDied","Data":"4634274d0596f237a6ff625c2528076f0e291464e7910506488e5f6db53eb461"} Oct 03 17:11:50 crc kubenswrapper[4744]: I1003 17:11:50.537662 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 03 17:11:51 crc kubenswrapper[4744]: I1003 17:11:51.489322 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmxd6" event={"ID":"68ad492e-af0b-4c93-9666-4f76715e8c2d","Type":"ContainerStarted","Data":"ce8392e91b74e87476886f3527eec6fc6deb50b1e6d416f3bdf6293484365aee"} Oct 03 17:11:51 crc kubenswrapper[4744]: I1003 17:11:51.513046 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qmxd6" podStartSLOduration=2.8630180320000003 podStartE2EDuration="5.513028536s" podCreationTimestamp="2025-10-03 17:11:46 +0000 UTC" firstStartedPulling="2025-10-03 17:11:48.4228422 +0000 UTC m=+2834.702718096" lastFinishedPulling="2025-10-03 17:11:51.072852704 +0000 UTC m=+2837.352728600" observedRunningTime="2025-10-03 17:11:51.503533168 +0000 UTC m=+2837.783409064" watchObservedRunningTime="2025-10-03 17:11:51.513028536 +0000 UTC m=+2837.792904432" Oct 03 17:11:52 crc kubenswrapper[4744]: I1003 17:11:52.505762 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"595f783e-db00-49b2-968f-79163001ffdd","Type":"ContainerStarted","Data":"81d520babfe8bb321dedaa86a51f6868965891969a4eba83624f528fedc8b468"} Oct 03 17:11:52 crc kubenswrapper[4744]: I1003 17:11:52.528535 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.038705251 podStartE2EDuration="49.528515656s" podCreationTimestamp="2025-10-03 17:11:03 +0000 UTC" firstStartedPulling="2025-10-03 17:11:05.044895302 +0000 UTC m=+2791.324771198" lastFinishedPulling="2025-10-03 17:11:50.534705707 +0000 UTC m=+2836.814581603" observedRunningTime="2025-10-03 17:11:52.522078635 +0000 UTC m=+2838.801954551" watchObservedRunningTime="2025-10-03 17:11:52.528515656 +0000 UTC m=+2838.808391552" Oct 03 17:11:57 crc kubenswrapper[4744]: I1003 17:11:57.319632 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:11:57 crc kubenswrapper[4744]: I1003 17:11:57.320627 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:11:57 crc kubenswrapper[4744]: I1003 17:11:57.392244 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:11:57 crc kubenswrapper[4744]: I1003 17:11:57.633227 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:11:57 crc kubenswrapper[4744]: I1003 17:11:57.689184 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qmxd6"] Oct 03 17:11:59 crc kubenswrapper[4744]: I1003 17:11:59.580287 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qmxd6" podUID="68ad492e-af0b-4c93-9666-4f76715e8c2d" containerName="registry-server" containerID="cri-o://ce8392e91b74e87476886f3527eec6fc6deb50b1e6d416f3bdf6293484365aee" gracePeriod=2 Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.031527 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.088548 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ad492e-af0b-4c93-9666-4f76715e8c2d-utilities\") pod \"68ad492e-af0b-4c93-9666-4f76715e8c2d\" (UID: \"68ad492e-af0b-4c93-9666-4f76715e8c2d\") " Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.088755 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ad492e-af0b-4c93-9666-4f76715e8c2d-catalog-content\") pod \"68ad492e-af0b-4c93-9666-4f76715e8c2d\" (UID: \"68ad492e-af0b-4c93-9666-4f76715e8c2d\") " Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.088828 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9gwf\" (UniqueName: \"kubernetes.io/projected/68ad492e-af0b-4c93-9666-4f76715e8c2d-kube-api-access-n9gwf\") pod \"68ad492e-af0b-4c93-9666-4f76715e8c2d\" (UID: \"68ad492e-af0b-4c93-9666-4f76715e8c2d\") " Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.090935 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68ad492e-af0b-4c93-9666-4f76715e8c2d-utilities" (OuterVolumeSpecName: "utilities") pod "68ad492e-af0b-4c93-9666-4f76715e8c2d" (UID: "68ad492e-af0b-4c93-9666-4f76715e8c2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.096235 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68ad492e-af0b-4c93-9666-4f76715e8c2d-kube-api-access-n9gwf" (OuterVolumeSpecName: "kube-api-access-n9gwf") pod "68ad492e-af0b-4c93-9666-4f76715e8c2d" (UID: "68ad492e-af0b-4c93-9666-4f76715e8c2d"). InnerVolumeSpecName "kube-api-access-n9gwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.102128 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68ad492e-af0b-4c93-9666-4f76715e8c2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68ad492e-af0b-4c93-9666-4f76715e8c2d" (UID: "68ad492e-af0b-4c93-9666-4f76715e8c2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.192199 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ad492e-af0b-4c93-9666-4f76715e8c2d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.192241 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9gwf\" (UniqueName: \"kubernetes.io/projected/68ad492e-af0b-4c93-9666-4f76715e8c2d-kube-api-access-n9gwf\") on node \"crc\" DevicePath \"\"" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.192257 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ad492e-af0b-4c93-9666-4f76715e8c2d-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.600424 4744 generic.go:334] "Generic (PLEG): container finished" podID="68ad492e-af0b-4c93-9666-4f76715e8c2d" containerID="ce8392e91b74e87476886f3527eec6fc6deb50b1e6d416f3bdf6293484365aee" exitCode=0 Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.600464 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmxd6" event={"ID":"68ad492e-af0b-4c93-9666-4f76715e8c2d","Type":"ContainerDied","Data":"ce8392e91b74e87476886f3527eec6fc6deb50b1e6d416f3bdf6293484365aee"} Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.600502 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qmxd6" event={"ID":"68ad492e-af0b-4c93-9666-4f76715e8c2d","Type":"ContainerDied","Data":"45568794ede4b8dcd6bcec127937814d01b12421a12361945d19ea260038bf6f"} Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.600519 4744 scope.go:117] "RemoveContainer" containerID="ce8392e91b74e87476886f3527eec6fc6deb50b1e6d416f3bdf6293484365aee" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.600634 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qmxd6" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.636361 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qmxd6"] Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.643026 4744 scope.go:117] "RemoveContainer" containerID="4634274d0596f237a6ff625c2528076f0e291464e7910506488e5f6db53eb461" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.645254 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qmxd6"] Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.680926 4744 scope.go:117] "RemoveContainer" containerID="3f1b8eecb5e4d879e58f295c2a35e8592d96ca33b3dc540d1fe2a6cbd2b0d705" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.732692 4744 scope.go:117] "RemoveContainer" containerID="ce8392e91b74e87476886f3527eec6fc6deb50b1e6d416f3bdf6293484365aee" Oct 03 17:12:00 crc kubenswrapper[4744]: E1003 17:12:00.733437 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce8392e91b74e87476886f3527eec6fc6deb50b1e6d416f3bdf6293484365aee\": container with ID starting with ce8392e91b74e87476886f3527eec6fc6deb50b1e6d416f3bdf6293484365aee not found: ID does not exist" containerID="ce8392e91b74e87476886f3527eec6fc6deb50b1e6d416f3bdf6293484365aee" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.733477 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce8392e91b74e87476886f3527eec6fc6deb50b1e6d416f3bdf6293484365aee"} err="failed to get container status \"ce8392e91b74e87476886f3527eec6fc6deb50b1e6d416f3bdf6293484365aee\": rpc error: code = NotFound desc = could not find container \"ce8392e91b74e87476886f3527eec6fc6deb50b1e6d416f3bdf6293484365aee\": container with ID starting with ce8392e91b74e87476886f3527eec6fc6deb50b1e6d416f3bdf6293484365aee not found: ID does not exist" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.733521 4744 scope.go:117] "RemoveContainer" containerID="4634274d0596f237a6ff625c2528076f0e291464e7910506488e5f6db53eb461" Oct 03 17:12:00 crc kubenswrapper[4744]: E1003 17:12:00.734018 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4634274d0596f237a6ff625c2528076f0e291464e7910506488e5f6db53eb461\": container with ID starting with 4634274d0596f237a6ff625c2528076f0e291464e7910506488e5f6db53eb461 not found: ID does not exist" containerID="4634274d0596f237a6ff625c2528076f0e291464e7910506488e5f6db53eb461" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.734069 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4634274d0596f237a6ff625c2528076f0e291464e7910506488e5f6db53eb461"} err="failed to get container status \"4634274d0596f237a6ff625c2528076f0e291464e7910506488e5f6db53eb461\": rpc error: code = NotFound desc = could not find container \"4634274d0596f237a6ff625c2528076f0e291464e7910506488e5f6db53eb461\": container with ID starting with 4634274d0596f237a6ff625c2528076f0e291464e7910506488e5f6db53eb461 not found: ID does not exist" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.734100 4744 scope.go:117] "RemoveContainer" containerID="3f1b8eecb5e4d879e58f295c2a35e8592d96ca33b3dc540d1fe2a6cbd2b0d705" Oct 03 17:12:00 crc kubenswrapper[4744]: E1003 17:12:00.734394 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f1b8eecb5e4d879e58f295c2a35e8592d96ca33b3dc540d1fe2a6cbd2b0d705\": container with ID starting with 3f1b8eecb5e4d879e58f295c2a35e8592d96ca33b3dc540d1fe2a6cbd2b0d705 not found: ID does not exist" containerID="3f1b8eecb5e4d879e58f295c2a35e8592d96ca33b3dc540d1fe2a6cbd2b0d705" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.734418 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f1b8eecb5e4d879e58f295c2a35e8592d96ca33b3dc540d1fe2a6cbd2b0d705"} err="failed to get container status \"3f1b8eecb5e4d879e58f295c2a35e8592d96ca33b3dc540d1fe2a6cbd2b0d705\": rpc error: code = NotFound desc = could not find container \"3f1b8eecb5e4d879e58f295c2a35e8592d96ca33b3dc540d1fe2a6cbd2b0d705\": container with ID starting with 3f1b8eecb5e4d879e58f295c2a35e8592d96ca33b3dc540d1fe2a6cbd2b0d705 not found: ID does not exist" Oct 03 17:12:00 crc kubenswrapper[4744]: I1003 17:12:00.905247 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68ad492e-af0b-4c93-9666-4f76715e8c2d" path="/var/lib/kubelet/pods/68ad492e-af0b-4c93-9666-4f76715e8c2d/volumes" Oct 03 17:12:04 crc kubenswrapper[4744]: I1003 17:12:04.668876 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:12:04 crc kubenswrapper[4744]: I1003 17:12:04.669360 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:12:04 crc kubenswrapper[4744]: I1003 17:12:04.669405 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 17:12:04 crc kubenswrapper[4744]: I1003 17:12:04.670227 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5e65036ce88f44b88c0f29775217f2811e8ec53114ae1711b6cfd674a3ffb8ab"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 17:12:04 crc kubenswrapper[4744]: I1003 17:12:04.670292 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://5e65036ce88f44b88c0f29775217f2811e8ec53114ae1711b6cfd674a3ffb8ab" gracePeriod=600 Oct 03 17:12:05 crc kubenswrapper[4744]: I1003 17:12:05.674720 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="5e65036ce88f44b88c0f29775217f2811e8ec53114ae1711b6cfd674a3ffb8ab" exitCode=0 Oct 03 17:12:05 crc kubenswrapper[4744]: I1003 17:12:05.674807 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"5e65036ce88f44b88c0f29775217f2811e8ec53114ae1711b6cfd674a3ffb8ab"} Oct 03 17:12:05 crc kubenswrapper[4744]: I1003 17:12:05.675385 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802"} Oct 03 17:12:05 crc kubenswrapper[4744]: I1003 17:12:05.675435 4744 scope.go:117] "RemoveContainer" containerID="31997b58df24ceb4691a7305f583e14ba328735ddb64d3cf3699ffa679b2e276" Oct 03 17:14:04 crc kubenswrapper[4744]: I1003 17:14:04.668785 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:14:04 crc kubenswrapper[4744]: I1003 17:14:04.669268 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:14:34 crc kubenswrapper[4744]: I1003 17:14:34.668450 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:14:34 crc kubenswrapper[4744]: I1003 17:14:34.669108 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.176351 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp"] Oct 03 17:15:00 crc kubenswrapper[4744]: E1003 17:15:00.177655 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ad492e-af0b-4c93-9666-4f76715e8c2d" containerName="registry-server" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.177668 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ad492e-af0b-4c93-9666-4f76715e8c2d" containerName="registry-server" Oct 03 17:15:00 crc kubenswrapper[4744]: E1003 17:15:00.177696 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ad492e-af0b-4c93-9666-4f76715e8c2d" containerName="extract-utilities" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.177704 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ad492e-af0b-4c93-9666-4f76715e8c2d" containerName="extract-utilities" Oct 03 17:15:00 crc kubenswrapper[4744]: E1003 17:15:00.177718 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ad492e-af0b-4c93-9666-4f76715e8c2d" containerName="extract-content" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.177725 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ad492e-af0b-4c93-9666-4f76715e8c2d" containerName="extract-content" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.177918 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="68ad492e-af0b-4c93-9666-4f76715e8c2d" containerName="registry-server" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.178669 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.181367 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.185329 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.211127 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8rnx\" (UniqueName: \"kubernetes.io/projected/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-kube-api-access-c8rnx\") pod \"collect-profiles-29325195-7ljgp\" (UID: \"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.211237 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-secret-volume\") pod \"collect-profiles-29325195-7ljgp\" (UID: \"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.211331 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-config-volume\") pod \"collect-profiles-29325195-7ljgp\" (UID: \"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.245161 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp"] Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.313676 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8rnx\" (UniqueName: \"kubernetes.io/projected/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-kube-api-access-c8rnx\") pod \"collect-profiles-29325195-7ljgp\" (UID: \"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.313779 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-secret-volume\") pod \"collect-profiles-29325195-7ljgp\" (UID: \"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.313896 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-config-volume\") pod \"collect-profiles-29325195-7ljgp\" (UID: \"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.315173 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-config-volume\") pod \"collect-profiles-29325195-7ljgp\" (UID: \"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.322741 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-secret-volume\") pod \"collect-profiles-29325195-7ljgp\" (UID: \"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.340149 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8rnx\" (UniqueName: \"kubernetes.io/projected/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-kube-api-access-c8rnx\") pod \"collect-profiles-29325195-7ljgp\" (UID: \"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" Oct 03 17:15:00 crc kubenswrapper[4744]: I1003 17:15:00.498136 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" Oct 03 17:15:01 crc kubenswrapper[4744]: I1003 17:15:01.418619 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp"] Oct 03 17:15:02 crc kubenswrapper[4744]: I1003 17:15:02.417232 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" event={"ID":"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b","Type":"ContainerStarted","Data":"706e631024b96f24af3381a83fd523097748542a7a4a251385063f9e7dcfb3fb"} Oct 03 17:15:02 crc kubenswrapper[4744]: I1003 17:15:02.417764 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" event={"ID":"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b","Type":"ContainerStarted","Data":"e940a1956b377ff92b36466383e008583ca034d7267249ecb9303df28d849e84"} Oct 03 17:15:02 crc kubenswrapper[4744]: I1003 17:15:02.436308 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" podStartSLOduration=2.436286341 podStartE2EDuration="2.436286341s" podCreationTimestamp="2025-10-03 17:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 17:15:02.434083596 +0000 UTC m=+3028.713959532" watchObservedRunningTime="2025-10-03 17:15:02.436286341 +0000 UTC m=+3028.716162237" Oct 03 17:15:03 crc kubenswrapper[4744]: I1003 17:15:03.427929 4744 generic.go:334] "Generic (PLEG): container finished" podID="04f0b7a4-1c87-4a84-bc35-24395c7e4b5b" containerID="706e631024b96f24af3381a83fd523097748542a7a4a251385063f9e7dcfb3fb" exitCode=0 Oct 03 17:15:03 crc kubenswrapper[4744]: I1003 17:15:03.427975 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" event={"ID":"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b","Type":"ContainerDied","Data":"706e631024b96f24af3381a83fd523097748542a7a4a251385063f9e7dcfb3fb"} Oct 03 17:15:04 crc kubenswrapper[4744]: I1003 17:15:04.668220 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:15:04 crc kubenswrapper[4744]: I1003 17:15:04.668891 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:15:04 crc kubenswrapper[4744]: I1003 17:15:04.670487 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 17:15:04 crc kubenswrapper[4744]: I1003 17:15:04.671265 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 17:15:04 crc kubenswrapper[4744]: I1003 17:15:04.671331 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" gracePeriod=600 Oct 03 17:15:04 crc kubenswrapper[4744]: E1003 17:15:04.802056 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.053161 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.142211 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8rnx\" (UniqueName: \"kubernetes.io/projected/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-kube-api-access-c8rnx\") pod \"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b\" (UID: \"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b\") " Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.142707 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-config-volume\") pod \"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b\" (UID: \"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b\") " Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.142948 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-secret-volume\") pod \"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b\" (UID: \"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b\") " Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.145149 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-config-volume" (OuterVolumeSpecName: "config-volume") pod "04f0b7a4-1c87-4a84-bc35-24395c7e4b5b" (UID: "04f0b7a4-1c87-4a84-bc35-24395c7e4b5b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.152735 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "04f0b7a4-1c87-4a84-bc35-24395c7e4b5b" (UID: "04f0b7a4-1c87-4a84-bc35-24395c7e4b5b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.153111 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-kube-api-access-c8rnx" (OuterVolumeSpecName: "kube-api-access-c8rnx") pod "04f0b7a4-1c87-4a84-bc35-24395c7e4b5b" (UID: "04f0b7a4-1c87-4a84-bc35-24395c7e4b5b"). InnerVolumeSpecName "kube-api-access-c8rnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.245347 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.245390 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8rnx\" (UniqueName: \"kubernetes.io/projected/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-kube-api-access-c8rnx\") on node \"crc\" DevicePath \"\"" Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.245399 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/04f0b7a4-1c87-4a84-bc35-24395c7e4b5b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.449298 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" event={"ID":"04f0b7a4-1c87-4a84-bc35-24395c7e4b5b","Type":"ContainerDied","Data":"e940a1956b377ff92b36466383e008583ca034d7267249ecb9303df28d849e84"} Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.449342 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325195-7ljgp" Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.449354 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e940a1956b377ff92b36466383e008583ca034d7267249ecb9303df28d849e84" Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.458015 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" exitCode=0 Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.458064 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802"} Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.458103 4744 scope.go:117] "RemoveContainer" containerID="5e65036ce88f44b88c0f29775217f2811e8ec53114ae1711b6cfd674a3ffb8ab" Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.459122 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:15:05 crc kubenswrapper[4744]: E1003 17:15:05.459531 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.525317 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt"] Oct 03 17:15:05 crc kubenswrapper[4744]: I1003 17:15:05.533664 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325150-4ktxt"] Oct 03 17:15:06 crc kubenswrapper[4744]: I1003 17:15:06.919303 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f" path="/var/lib/kubelet/pods/bf720b18-16f4-4bd9-b2ea-14ceca6f8b5f/volumes" Oct 03 17:15:15 crc kubenswrapper[4744]: I1003 17:15:15.891876 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:15:15 crc kubenswrapper[4744]: E1003 17:15:15.892663 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:15:28 crc kubenswrapper[4744]: I1003 17:15:28.892598 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:15:28 crc kubenswrapper[4744]: E1003 17:15:28.893989 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:15:40 crc kubenswrapper[4744]: I1003 17:15:40.263339 4744 scope.go:117] "RemoveContainer" containerID="460cfb11a0f81225424d3d509a800349764ca50030e40fbccac077bce1c5a4fc" Oct 03 17:15:41 crc kubenswrapper[4744]: I1003 17:15:41.893837 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:15:41 crc kubenswrapper[4744]: E1003 17:15:41.894386 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:15:54 crc kubenswrapper[4744]: I1003 17:15:54.906536 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:15:54 crc kubenswrapper[4744]: E1003 17:15:54.907318 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:16:07 crc kubenswrapper[4744]: I1003 17:16:07.893097 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:16:07 crc kubenswrapper[4744]: E1003 17:16:07.894033 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:16:18 crc kubenswrapper[4744]: I1003 17:16:18.892470 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:16:18 crc kubenswrapper[4744]: E1003 17:16:18.893263 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:16:32 crc kubenswrapper[4744]: I1003 17:16:32.892184 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:16:32 crc kubenswrapper[4744]: E1003 17:16:32.893451 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:16:47 crc kubenswrapper[4744]: I1003 17:16:47.892736 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:16:47 crc kubenswrapper[4744]: E1003 17:16:47.893633 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:17:02 crc kubenswrapper[4744]: I1003 17:17:02.892531 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:17:02 crc kubenswrapper[4744]: E1003 17:17:02.893345 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:17:16 crc kubenswrapper[4744]: I1003 17:17:16.892259 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:17:16 crc kubenswrapper[4744]: E1003 17:17:16.892952 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:17:29 crc kubenswrapper[4744]: I1003 17:17:29.892312 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:17:29 crc kubenswrapper[4744]: E1003 17:17:29.894409 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:17:40 crc kubenswrapper[4744]: I1003 17:17:40.891924 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:17:40 crc kubenswrapper[4744]: E1003 17:17:40.892783 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:17:54 crc kubenswrapper[4744]: I1003 17:17:54.914607 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:17:54 crc kubenswrapper[4744]: E1003 17:17:54.915722 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:18:07 crc kubenswrapper[4744]: I1003 17:18:07.893093 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:18:07 crc kubenswrapper[4744]: E1003 17:18:07.893887 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:18:18 crc kubenswrapper[4744]: I1003 17:18:18.892626 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:18:18 crc kubenswrapper[4744]: E1003 17:18:18.893767 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:18:27 crc kubenswrapper[4744]: I1003 17:18:27.898454 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6hjkm"] Oct 03 17:18:27 crc kubenswrapper[4744]: E1003 17:18:27.900294 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04f0b7a4-1c87-4a84-bc35-24395c7e4b5b" containerName="collect-profiles" Oct 03 17:18:27 crc kubenswrapper[4744]: I1003 17:18:27.900321 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="04f0b7a4-1c87-4a84-bc35-24395c7e4b5b" containerName="collect-profiles" Oct 03 17:18:27 crc kubenswrapper[4744]: I1003 17:18:27.900693 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="04f0b7a4-1c87-4a84-bc35-24395c7e4b5b" containerName="collect-profiles" Oct 03 17:18:27 crc kubenswrapper[4744]: I1003 17:18:27.910075 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6hjkm"] Oct 03 17:18:27 crc kubenswrapper[4744]: I1003 17:18:27.910204 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:27 crc kubenswrapper[4744]: I1003 17:18:27.997413 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b19435b-47d0-4323-ad41-bb0dd887192a-catalog-content\") pod \"community-operators-6hjkm\" (UID: \"9b19435b-47d0-4323-ad41-bb0dd887192a\") " pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:27 crc kubenswrapper[4744]: I1003 17:18:27.997571 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b19435b-47d0-4323-ad41-bb0dd887192a-utilities\") pod \"community-operators-6hjkm\" (UID: \"9b19435b-47d0-4323-ad41-bb0dd887192a\") " pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:27 crc kubenswrapper[4744]: I1003 17:18:27.997763 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgg2d\" (UniqueName: \"kubernetes.io/projected/9b19435b-47d0-4323-ad41-bb0dd887192a-kube-api-access-cgg2d\") pod \"community-operators-6hjkm\" (UID: \"9b19435b-47d0-4323-ad41-bb0dd887192a\") " pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:28 crc kubenswrapper[4744]: I1003 17:18:28.099844 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b19435b-47d0-4323-ad41-bb0dd887192a-catalog-content\") pod \"community-operators-6hjkm\" (UID: \"9b19435b-47d0-4323-ad41-bb0dd887192a\") " pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:28 crc kubenswrapper[4744]: I1003 17:18:28.099936 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b19435b-47d0-4323-ad41-bb0dd887192a-utilities\") pod \"community-operators-6hjkm\" (UID: \"9b19435b-47d0-4323-ad41-bb0dd887192a\") " pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:28 crc kubenswrapper[4744]: I1003 17:18:28.100002 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgg2d\" (UniqueName: \"kubernetes.io/projected/9b19435b-47d0-4323-ad41-bb0dd887192a-kube-api-access-cgg2d\") pod \"community-operators-6hjkm\" (UID: \"9b19435b-47d0-4323-ad41-bb0dd887192a\") " pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:28 crc kubenswrapper[4744]: I1003 17:18:28.100681 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b19435b-47d0-4323-ad41-bb0dd887192a-catalog-content\") pod \"community-operators-6hjkm\" (UID: \"9b19435b-47d0-4323-ad41-bb0dd887192a\") " pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:28 crc kubenswrapper[4744]: I1003 17:18:28.100905 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b19435b-47d0-4323-ad41-bb0dd887192a-utilities\") pod \"community-operators-6hjkm\" (UID: \"9b19435b-47d0-4323-ad41-bb0dd887192a\") " pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:28 crc kubenswrapper[4744]: I1003 17:18:28.118594 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgg2d\" (UniqueName: \"kubernetes.io/projected/9b19435b-47d0-4323-ad41-bb0dd887192a-kube-api-access-cgg2d\") pod \"community-operators-6hjkm\" (UID: \"9b19435b-47d0-4323-ad41-bb0dd887192a\") " pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:28 crc kubenswrapper[4744]: I1003 17:18:28.257338 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:28 crc kubenswrapper[4744]: I1003 17:18:28.797784 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6hjkm"] Oct 03 17:18:29 crc kubenswrapper[4744]: I1003 17:18:29.424446 4744 generic.go:334] "Generic (PLEG): container finished" podID="9b19435b-47d0-4323-ad41-bb0dd887192a" containerID="14cd113b4ebea2986e7e4dc21e489e912a4f70b24ca3c20c90551973c1744631" exitCode=0 Oct 03 17:18:29 crc kubenswrapper[4744]: I1003 17:18:29.424493 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6hjkm" event={"ID":"9b19435b-47d0-4323-ad41-bb0dd887192a","Type":"ContainerDied","Data":"14cd113b4ebea2986e7e4dc21e489e912a4f70b24ca3c20c90551973c1744631"} Oct 03 17:18:29 crc kubenswrapper[4744]: I1003 17:18:29.424580 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6hjkm" event={"ID":"9b19435b-47d0-4323-ad41-bb0dd887192a","Type":"ContainerStarted","Data":"0eda8513b6032fc97812be094bb87cb0eb8da516a6f0b0c0cffa2294eaac45cd"} Oct 03 17:18:29 crc kubenswrapper[4744]: I1003 17:18:29.426251 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 17:18:30 crc kubenswrapper[4744]: I1003 17:18:30.437472 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6hjkm" event={"ID":"9b19435b-47d0-4323-ad41-bb0dd887192a","Type":"ContainerStarted","Data":"302394ee6f5616c04bcaa29d252a8006f9b37d640bcac53e2a04845eda477c8b"} Oct 03 17:18:33 crc kubenswrapper[4744]: I1003 17:18:33.466930 4744 generic.go:334] "Generic (PLEG): container finished" podID="9b19435b-47d0-4323-ad41-bb0dd887192a" containerID="302394ee6f5616c04bcaa29d252a8006f9b37d640bcac53e2a04845eda477c8b" exitCode=0 Oct 03 17:18:33 crc kubenswrapper[4744]: I1003 17:18:33.467016 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6hjkm" event={"ID":"9b19435b-47d0-4323-ad41-bb0dd887192a","Type":"ContainerDied","Data":"302394ee6f5616c04bcaa29d252a8006f9b37d640bcac53e2a04845eda477c8b"} Oct 03 17:18:33 crc kubenswrapper[4744]: I1003 17:18:33.892626 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:18:33 crc kubenswrapper[4744]: E1003 17:18:33.893166 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:18:34 crc kubenswrapper[4744]: I1003 17:18:34.478505 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6hjkm" event={"ID":"9b19435b-47d0-4323-ad41-bb0dd887192a","Type":"ContainerStarted","Data":"1a2e6b4af1fac0f6a01f96a1fe70f0487f8494eeb65505dcacc314ea9e7fedaa"} Oct 03 17:18:34 crc kubenswrapper[4744]: I1003 17:18:34.502191 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6hjkm" podStartSLOduration=2.716767486 podStartE2EDuration="7.502168147s" podCreationTimestamp="2025-10-03 17:18:27 +0000 UTC" firstStartedPulling="2025-10-03 17:18:29.426004226 +0000 UTC m=+3235.705880122" lastFinishedPulling="2025-10-03 17:18:34.211404887 +0000 UTC m=+3240.491280783" observedRunningTime="2025-10-03 17:18:34.496631808 +0000 UTC m=+3240.776507724" watchObservedRunningTime="2025-10-03 17:18:34.502168147 +0000 UTC m=+3240.782044053" Oct 03 17:18:38 crc kubenswrapper[4744]: I1003 17:18:38.257621 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:38 crc kubenswrapper[4744]: I1003 17:18:38.258749 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:38 crc kubenswrapper[4744]: I1003 17:18:38.305183 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:39 crc kubenswrapper[4744]: I1003 17:18:39.587695 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:39 crc kubenswrapper[4744]: I1003 17:18:39.647528 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6hjkm"] Oct 03 17:18:41 crc kubenswrapper[4744]: I1003 17:18:41.543059 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6hjkm" podUID="9b19435b-47d0-4323-ad41-bb0dd887192a" containerName="registry-server" containerID="cri-o://1a2e6b4af1fac0f6a01f96a1fe70f0487f8494eeb65505dcacc314ea9e7fedaa" gracePeriod=2 Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.231275 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.304920 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b19435b-47d0-4323-ad41-bb0dd887192a-utilities\") pod \"9b19435b-47d0-4323-ad41-bb0dd887192a\" (UID: \"9b19435b-47d0-4323-ad41-bb0dd887192a\") " Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.305086 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b19435b-47d0-4323-ad41-bb0dd887192a-catalog-content\") pod \"9b19435b-47d0-4323-ad41-bb0dd887192a\" (UID: \"9b19435b-47d0-4323-ad41-bb0dd887192a\") " Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.305131 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgg2d\" (UniqueName: \"kubernetes.io/projected/9b19435b-47d0-4323-ad41-bb0dd887192a-kube-api-access-cgg2d\") pod \"9b19435b-47d0-4323-ad41-bb0dd887192a\" (UID: \"9b19435b-47d0-4323-ad41-bb0dd887192a\") " Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.305809 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b19435b-47d0-4323-ad41-bb0dd887192a-utilities" (OuterVolumeSpecName: "utilities") pod "9b19435b-47d0-4323-ad41-bb0dd887192a" (UID: "9b19435b-47d0-4323-ad41-bb0dd887192a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.323814 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b19435b-47d0-4323-ad41-bb0dd887192a-kube-api-access-cgg2d" (OuterVolumeSpecName: "kube-api-access-cgg2d") pod "9b19435b-47d0-4323-ad41-bb0dd887192a" (UID: "9b19435b-47d0-4323-ad41-bb0dd887192a"). InnerVolumeSpecName "kube-api-access-cgg2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.357088 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b19435b-47d0-4323-ad41-bb0dd887192a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9b19435b-47d0-4323-ad41-bb0dd887192a" (UID: "9b19435b-47d0-4323-ad41-bb0dd887192a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.408070 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b19435b-47d0-4323-ad41-bb0dd887192a-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.408111 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b19435b-47d0-4323-ad41-bb0dd887192a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.408127 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgg2d\" (UniqueName: \"kubernetes.io/projected/9b19435b-47d0-4323-ad41-bb0dd887192a-kube-api-access-cgg2d\") on node \"crc\" DevicePath \"\"" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.554259 4744 generic.go:334] "Generic (PLEG): container finished" podID="9b19435b-47d0-4323-ad41-bb0dd887192a" containerID="1a2e6b4af1fac0f6a01f96a1fe70f0487f8494eeb65505dcacc314ea9e7fedaa" exitCode=0 Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.554315 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6hjkm" event={"ID":"9b19435b-47d0-4323-ad41-bb0dd887192a","Type":"ContainerDied","Data":"1a2e6b4af1fac0f6a01f96a1fe70f0487f8494eeb65505dcacc314ea9e7fedaa"} Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.554333 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6hjkm" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.554347 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6hjkm" event={"ID":"9b19435b-47d0-4323-ad41-bb0dd887192a","Type":"ContainerDied","Data":"0eda8513b6032fc97812be094bb87cb0eb8da516a6f0b0c0cffa2294eaac45cd"} Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.554373 4744 scope.go:117] "RemoveContainer" containerID="1a2e6b4af1fac0f6a01f96a1fe70f0487f8494eeb65505dcacc314ea9e7fedaa" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.578220 4744 scope.go:117] "RemoveContainer" containerID="302394ee6f5616c04bcaa29d252a8006f9b37d640bcac53e2a04845eda477c8b" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.601139 4744 scope.go:117] "RemoveContainer" containerID="14cd113b4ebea2986e7e4dc21e489e912a4f70b24ca3c20c90551973c1744631" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.606242 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6hjkm"] Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.615191 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6hjkm"] Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.667394 4744 scope.go:117] "RemoveContainer" containerID="1a2e6b4af1fac0f6a01f96a1fe70f0487f8494eeb65505dcacc314ea9e7fedaa" Oct 03 17:18:42 crc kubenswrapper[4744]: E1003 17:18:42.667962 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a2e6b4af1fac0f6a01f96a1fe70f0487f8494eeb65505dcacc314ea9e7fedaa\": container with ID starting with 1a2e6b4af1fac0f6a01f96a1fe70f0487f8494eeb65505dcacc314ea9e7fedaa not found: ID does not exist" containerID="1a2e6b4af1fac0f6a01f96a1fe70f0487f8494eeb65505dcacc314ea9e7fedaa" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.667991 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a2e6b4af1fac0f6a01f96a1fe70f0487f8494eeb65505dcacc314ea9e7fedaa"} err="failed to get container status \"1a2e6b4af1fac0f6a01f96a1fe70f0487f8494eeb65505dcacc314ea9e7fedaa\": rpc error: code = NotFound desc = could not find container \"1a2e6b4af1fac0f6a01f96a1fe70f0487f8494eeb65505dcacc314ea9e7fedaa\": container with ID starting with 1a2e6b4af1fac0f6a01f96a1fe70f0487f8494eeb65505dcacc314ea9e7fedaa not found: ID does not exist" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.668012 4744 scope.go:117] "RemoveContainer" containerID="302394ee6f5616c04bcaa29d252a8006f9b37d640bcac53e2a04845eda477c8b" Oct 03 17:18:42 crc kubenswrapper[4744]: E1003 17:18:42.668343 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"302394ee6f5616c04bcaa29d252a8006f9b37d640bcac53e2a04845eda477c8b\": container with ID starting with 302394ee6f5616c04bcaa29d252a8006f9b37d640bcac53e2a04845eda477c8b not found: ID does not exist" containerID="302394ee6f5616c04bcaa29d252a8006f9b37d640bcac53e2a04845eda477c8b" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.668394 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"302394ee6f5616c04bcaa29d252a8006f9b37d640bcac53e2a04845eda477c8b"} err="failed to get container status \"302394ee6f5616c04bcaa29d252a8006f9b37d640bcac53e2a04845eda477c8b\": rpc error: code = NotFound desc = could not find container \"302394ee6f5616c04bcaa29d252a8006f9b37d640bcac53e2a04845eda477c8b\": container with ID starting with 302394ee6f5616c04bcaa29d252a8006f9b37d640bcac53e2a04845eda477c8b not found: ID does not exist" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.668421 4744 scope.go:117] "RemoveContainer" containerID="14cd113b4ebea2986e7e4dc21e489e912a4f70b24ca3c20c90551973c1744631" Oct 03 17:18:42 crc kubenswrapper[4744]: E1003 17:18:42.668939 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14cd113b4ebea2986e7e4dc21e489e912a4f70b24ca3c20c90551973c1744631\": container with ID starting with 14cd113b4ebea2986e7e4dc21e489e912a4f70b24ca3c20c90551973c1744631 not found: ID does not exist" containerID="14cd113b4ebea2986e7e4dc21e489e912a4f70b24ca3c20c90551973c1744631" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.668962 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14cd113b4ebea2986e7e4dc21e489e912a4f70b24ca3c20c90551973c1744631"} err="failed to get container status \"14cd113b4ebea2986e7e4dc21e489e912a4f70b24ca3c20c90551973c1744631\": rpc error: code = NotFound desc = could not find container \"14cd113b4ebea2986e7e4dc21e489e912a4f70b24ca3c20c90551973c1744631\": container with ID starting with 14cd113b4ebea2986e7e4dc21e489e912a4f70b24ca3c20c90551973c1744631 not found: ID does not exist" Oct 03 17:18:42 crc kubenswrapper[4744]: I1003 17:18:42.903453 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b19435b-47d0-4323-ad41-bb0dd887192a" path="/var/lib/kubelet/pods/9b19435b-47d0-4323-ad41-bb0dd887192a/volumes" Oct 03 17:18:45 crc kubenswrapper[4744]: I1003 17:18:45.891943 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:18:45 crc kubenswrapper[4744]: E1003 17:18:45.893259 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:18:59 crc kubenswrapper[4744]: I1003 17:18:59.891575 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:18:59 crc kubenswrapper[4744]: E1003 17:18:59.892391 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:19:14 crc kubenswrapper[4744]: I1003 17:19:14.898668 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:19:14 crc kubenswrapper[4744]: E1003 17:19:14.899294 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:19:25 crc kubenswrapper[4744]: I1003 17:19:25.892366 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:19:25 crc kubenswrapper[4744]: E1003 17:19:25.893189 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:19:39 crc kubenswrapper[4744]: I1003 17:19:39.892404 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:19:39 crc kubenswrapper[4744]: E1003 17:19:39.893210 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:19:50 crc kubenswrapper[4744]: I1003 17:19:50.895999 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:19:50 crc kubenswrapper[4744]: E1003 17:19:50.896990 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:20:03 crc kubenswrapper[4744]: I1003 17:20:03.892404 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:20:03 crc kubenswrapper[4744]: E1003 17:20:03.893281 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:20:17 crc kubenswrapper[4744]: I1003 17:20:17.892050 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:20:18 crc kubenswrapper[4744]: I1003 17:20:18.457589 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"cb4b2894f5fe1e71c4e0469f8b5747a451c56377648ebf3a4e30531cfe0daad2"} Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.134746 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2rm9d"] Oct 03 17:20:40 crc kubenswrapper[4744]: E1003 17:20:40.135524 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b19435b-47d0-4323-ad41-bb0dd887192a" containerName="extract-content" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.135536 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b19435b-47d0-4323-ad41-bb0dd887192a" containerName="extract-content" Oct 03 17:20:40 crc kubenswrapper[4744]: E1003 17:20:40.135544 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b19435b-47d0-4323-ad41-bb0dd887192a" containerName="extract-utilities" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.135549 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b19435b-47d0-4323-ad41-bb0dd887192a" containerName="extract-utilities" Oct 03 17:20:40 crc kubenswrapper[4744]: E1003 17:20:40.135570 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b19435b-47d0-4323-ad41-bb0dd887192a" containerName="registry-server" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.135576 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b19435b-47d0-4323-ad41-bb0dd887192a" containerName="registry-server" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.135772 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b19435b-47d0-4323-ad41-bb0dd887192a" containerName="registry-server" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.137056 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.146954 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2rm9d"] Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.206475 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkh7p\" (UniqueName: \"kubernetes.io/projected/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-kube-api-access-xkh7p\") pod \"redhat-operators-2rm9d\" (UID: \"5508746c-8415-44ca-aefe-2c8a6ca1c8e5\") " pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.206602 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-catalog-content\") pod \"redhat-operators-2rm9d\" (UID: \"5508746c-8415-44ca-aefe-2c8a6ca1c8e5\") " pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.206848 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-utilities\") pod \"redhat-operators-2rm9d\" (UID: \"5508746c-8415-44ca-aefe-2c8a6ca1c8e5\") " pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.308928 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkh7p\" (UniqueName: \"kubernetes.io/projected/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-kube-api-access-xkh7p\") pod \"redhat-operators-2rm9d\" (UID: \"5508746c-8415-44ca-aefe-2c8a6ca1c8e5\") " pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.308986 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-catalog-content\") pod \"redhat-operators-2rm9d\" (UID: \"5508746c-8415-44ca-aefe-2c8a6ca1c8e5\") " pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.309048 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-utilities\") pod \"redhat-operators-2rm9d\" (UID: \"5508746c-8415-44ca-aefe-2c8a6ca1c8e5\") " pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.309480 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-utilities\") pod \"redhat-operators-2rm9d\" (UID: \"5508746c-8415-44ca-aefe-2c8a6ca1c8e5\") " pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.309648 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-catalog-content\") pod \"redhat-operators-2rm9d\" (UID: \"5508746c-8415-44ca-aefe-2c8a6ca1c8e5\") " pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.337933 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkh7p\" (UniqueName: \"kubernetes.io/projected/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-kube-api-access-xkh7p\") pod \"redhat-operators-2rm9d\" (UID: \"5508746c-8415-44ca-aefe-2c8a6ca1c8e5\") " pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.470412 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:20:40 crc kubenswrapper[4744]: I1003 17:20:40.949251 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2rm9d"] Oct 03 17:20:41 crc kubenswrapper[4744]: I1003 17:20:41.664745 4744 generic.go:334] "Generic (PLEG): container finished" podID="5508746c-8415-44ca-aefe-2c8a6ca1c8e5" containerID="3b254c5a48f7f73bd008d26a9921365412ee3e9882b2adcdce764251daa19621" exitCode=0 Oct 03 17:20:41 crc kubenswrapper[4744]: I1003 17:20:41.664856 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rm9d" event={"ID":"5508746c-8415-44ca-aefe-2c8a6ca1c8e5","Type":"ContainerDied","Data":"3b254c5a48f7f73bd008d26a9921365412ee3e9882b2adcdce764251daa19621"} Oct 03 17:20:41 crc kubenswrapper[4744]: I1003 17:20:41.665145 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rm9d" event={"ID":"5508746c-8415-44ca-aefe-2c8a6ca1c8e5","Type":"ContainerStarted","Data":"7fad077279258a3ec2f6e40a49246815f596d6e39250070284ee7f0d279d5171"} Oct 03 17:20:43 crc kubenswrapper[4744]: I1003 17:20:43.682663 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rm9d" event={"ID":"5508746c-8415-44ca-aefe-2c8a6ca1c8e5","Type":"ContainerStarted","Data":"deeaf0240d132b4da0025f3fdbaa7a1904e180007768c6fc1b4b6272acb398bf"} Oct 03 17:20:46 crc kubenswrapper[4744]: I1003 17:20:46.714426 4744 generic.go:334] "Generic (PLEG): container finished" podID="5508746c-8415-44ca-aefe-2c8a6ca1c8e5" containerID="deeaf0240d132b4da0025f3fdbaa7a1904e180007768c6fc1b4b6272acb398bf" exitCode=0 Oct 03 17:20:46 crc kubenswrapper[4744]: I1003 17:20:46.714523 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rm9d" event={"ID":"5508746c-8415-44ca-aefe-2c8a6ca1c8e5","Type":"ContainerDied","Data":"deeaf0240d132b4da0025f3fdbaa7a1904e180007768c6fc1b4b6272acb398bf"} Oct 03 17:20:48 crc kubenswrapper[4744]: I1003 17:20:48.735789 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rm9d" event={"ID":"5508746c-8415-44ca-aefe-2c8a6ca1c8e5","Type":"ContainerStarted","Data":"5b2ec37f4bbc81a729603cfe6fccab7004eafc5fcafc417afb40391f4d754d2e"} Oct 03 17:20:48 crc kubenswrapper[4744]: I1003 17:20:48.755093 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2rm9d" podStartSLOduration=2.792035752 podStartE2EDuration="8.755075258s" podCreationTimestamp="2025-10-03 17:20:40 +0000 UTC" firstStartedPulling="2025-10-03 17:20:41.666697929 +0000 UTC m=+3367.946573825" lastFinishedPulling="2025-10-03 17:20:47.629737425 +0000 UTC m=+3373.909613331" observedRunningTime="2025-10-03 17:20:48.754011621 +0000 UTC m=+3375.033887527" watchObservedRunningTime="2025-10-03 17:20:48.755075258 +0000 UTC m=+3375.034951154" Oct 03 17:20:50 crc kubenswrapper[4744]: I1003 17:20:50.470710 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:20:50 crc kubenswrapper[4744]: I1003 17:20:50.471081 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:20:51 crc kubenswrapper[4744]: I1003 17:20:51.526551 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2rm9d" podUID="5508746c-8415-44ca-aefe-2c8a6ca1c8e5" containerName="registry-server" probeResult="failure" output=< Oct 03 17:20:51 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 03 17:20:51 crc kubenswrapper[4744]: > Oct 03 17:21:00 crc kubenswrapper[4744]: I1003 17:21:00.524003 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:21:00 crc kubenswrapper[4744]: I1003 17:21:00.574107 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:21:00 crc kubenswrapper[4744]: I1003 17:21:00.757934 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2rm9d"] Oct 03 17:21:01 crc kubenswrapper[4744]: I1003 17:21:01.850628 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2rm9d" podUID="5508746c-8415-44ca-aefe-2c8a6ca1c8e5" containerName="registry-server" containerID="cri-o://5b2ec37f4bbc81a729603cfe6fccab7004eafc5fcafc417afb40391f4d754d2e" gracePeriod=2 Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:02.861707 4744 generic.go:334] "Generic (PLEG): container finished" podID="5508746c-8415-44ca-aefe-2c8a6ca1c8e5" containerID="5b2ec37f4bbc81a729603cfe6fccab7004eafc5fcafc417afb40391f4d754d2e" exitCode=0 Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:02.861789 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rm9d" event={"ID":"5508746c-8415-44ca-aefe-2c8a6ca1c8e5","Type":"ContainerDied","Data":"5b2ec37f4bbc81a729603cfe6fccab7004eafc5fcafc417afb40391f4d754d2e"} Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.016212 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.135944 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-catalog-content\") pod \"5508746c-8415-44ca-aefe-2c8a6ca1c8e5\" (UID: \"5508746c-8415-44ca-aefe-2c8a6ca1c8e5\") " Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.136201 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkh7p\" (UniqueName: \"kubernetes.io/projected/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-kube-api-access-xkh7p\") pod \"5508746c-8415-44ca-aefe-2c8a6ca1c8e5\" (UID: \"5508746c-8415-44ca-aefe-2c8a6ca1c8e5\") " Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.136358 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-utilities\") pod \"5508746c-8415-44ca-aefe-2c8a6ca1c8e5\" (UID: \"5508746c-8415-44ca-aefe-2c8a6ca1c8e5\") " Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.137155 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-utilities" (OuterVolumeSpecName: "utilities") pod "5508746c-8415-44ca-aefe-2c8a6ca1c8e5" (UID: "5508746c-8415-44ca-aefe-2c8a6ca1c8e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.145937 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-kube-api-access-xkh7p" (OuterVolumeSpecName: "kube-api-access-xkh7p") pod "5508746c-8415-44ca-aefe-2c8a6ca1c8e5" (UID: "5508746c-8415-44ca-aefe-2c8a6ca1c8e5"). InnerVolumeSpecName "kube-api-access-xkh7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.235841 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5508746c-8415-44ca-aefe-2c8a6ca1c8e5" (UID: "5508746c-8415-44ca-aefe-2c8a6ca1c8e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.238476 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.238525 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.238538 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkh7p\" (UniqueName: \"kubernetes.io/projected/5508746c-8415-44ca-aefe-2c8a6ca1c8e5-kube-api-access-xkh7p\") on node \"crc\" DevicePath \"\"" Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.875689 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2rm9d" event={"ID":"5508746c-8415-44ca-aefe-2c8a6ca1c8e5","Type":"ContainerDied","Data":"7fad077279258a3ec2f6e40a49246815f596d6e39250070284ee7f0d279d5171"} Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.875789 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2rm9d" Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.876095 4744 scope.go:117] "RemoveContainer" containerID="5b2ec37f4bbc81a729603cfe6fccab7004eafc5fcafc417afb40391f4d754d2e" Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.931252 4744 scope.go:117] "RemoveContainer" containerID="deeaf0240d132b4da0025f3fdbaa7a1904e180007768c6fc1b4b6272acb398bf" Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.965551 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2rm9d"] Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.969702 4744 scope.go:117] "RemoveContainer" containerID="3b254c5a48f7f73bd008d26a9921365412ee3e9882b2adcdce764251daa19621" Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:03.973791 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2rm9d"] Oct 03 17:21:04 crc kubenswrapper[4744]: I1003 17:21:04.908681 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5508746c-8415-44ca-aefe-2c8a6ca1c8e5" path="/var/lib/kubelet/pods/5508746c-8415-44ca-aefe-2c8a6ca1c8e5/volumes" Oct 03 17:22:34 crc kubenswrapper[4744]: I1003 17:22:34.668142 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:22:34 crc kubenswrapper[4744]: I1003 17:22:34.668738 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.483396 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vb729"] Oct 03 17:22:42 crc kubenswrapper[4744]: E1003 17:22:42.484452 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5508746c-8415-44ca-aefe-2c8a6ca1c8e5" containerName="registry-server" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.484470 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5508746c-8415-44ca-aefe-2c8a6ca1c8e5" containerName="registry-server" Oct 03 17:22:42 crc kubenswrapper[4744]: E1003 17:22:42.484490 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5508746c-8415-44ca-aefe-2c8a6ca1c8e5" containerName="extract-content" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.484519 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5508746c-8415-44ca-aefe-2c8a6ca1c8e5" containerName="extract-content" Oct 03 17:22:42 crc kubenswrapper[4744]: E1003 17:22:42.484556 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5508746c-8415-44ca-aefe-2c8a6ca1c8e5" containerName="extract-utilities" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.484565 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5508746c-8415-44ca-aefe-2c8a6ca1c8e5" containerName="extract-utilities" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.484805 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5508746c-8415-44ca-aefe-2c8a6ca1c8e5" containerName="registry-server" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.486520 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.494859 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vb729"] Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.551694 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/082b714a-fdd1-4c11-9759-1c826bbab583-utilities\") pod \"redhat-marketplace-vb729\" (UID: \"082b714a-fdd1-4c11-9759-1c826bbab583\") " pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.551911 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/082b714a-fdd1-4c11-9759-1c826bbab583-catalog-content\") pod \"redhat-marketplace-vb729\" (UID: \"082b714a-fdd1-4c11-9759-1c826bbab583\") " pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.552017 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdzxt\" (UniqueName: \"kubernetes.io/projected/082b714a-fdd1-4c11-9759-1c826bbab583-kube-api-access-pdzxt\") pod \"redhat-marketplace-vb729\" (UID: \"082b714a-fdd1-4c11-9759-1c826bbab583\") " pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.653842 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdzxt\" (UniqueName: \"kubernetes.io/projected/082b714a-fdd1-4c11-9759-1c826bbab583-kube-api-access-pdzxt\") pod \"redhat-marketplace-vb729\" (UID: \"082b714a-fdd1-4c11-9759-1c826bbab583\") " pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.653940 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/082b714a-fdd1-4c11-9759-1c826bbab583-utilities\") pod \"redhat-marketplace-vb729\" (UID: \"082b714a-fdd1-4c11-9759-1c826bbab583\") " pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.654101 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/082b714a-fdd1-4c11-9759-1c826bbab583-catalog-content\") pod \"redhat-marketplace-vb729\" (UID: \"082b714a-fdd1-4c11-9759-1c826bbab583\") " pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.654744 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/082b714a-fdd1-4c11-9759-1c826bbab583-utilities\") pod \"redhat-marketplace-vb729\" (UID: \"082b714a-fdd1-4c11-9759-1c826bbab583\") " pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.654795 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/082b714a-fdd1-4c11-9759-1c826bbab583-catalog-content\") pod \"redhat-marketplace-vb729\" (UID: \"082b714a-fdd1-4c11-9759-1c826bbab583\") " pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.676716 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdzxt\" (UniqueName: \"kubernetes.io/projected/082b714a-fdd1-4c11-9759-1c826bbab583-kube-api-access-pdzxt\") pod \"redhat-marketplace-vb729\" (UID: \"082b714a-fdd1-4c11-9759-1c826bbab583\") " pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:42 crc kubenswrapper[4744]: I1003 17:22:42.812120 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:43 crc kubenswrapper[4744]: I1003 17:22:43.257785 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vb729"] Oct 03 17:22:43 crc kubenswrapper[4744]: I1003 17:22:43.808050 4744 generic.go:334] "Generic (PLEG): container finished" podID="082b714a-fdd1-4c11-9759-1c826bbab583" containerID="5dce479679a12a2b14a8e82cfd87936a8710f94392af5d67734c36fa4542c285" exitCode=0 Oct 03 17:22:43 crc kubenswrapper[4744]: I1003 17:22:43.808143 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vb729" event={"ID":"082b714a-fdd1-4c11-9759-1c826bbab583","Type":"ContainerDied","Data":"5dce479679a12a2b14a8e82cfd87936a8710f94392af5d67734c36fa4542c285"} Oct 03 17:22:43 crc kubenswrapper[4744]: I1003 17:22:43.808348 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vb729" event={"ID":"082b714a-fdd1-4c11-9759-1c826bbab583","Type":"ContainerStarted","Data":"77291cc5cc35327a1420cc99d2430e3b3e869246cff6f94a2c21c1d8a44049e8"} Oct 03 17:22:45 crc kubenswrapper[4744]: I1003 17:22:45.833089 4744 generic.go:334] "Generic (PLEG): container finished" podID="082b714a-fdd1-4c11-9759-1c826bbab583" containerID="2e878649e63a60755f8f3b49a1a3c4c3989bf77069d87ca1080f58d551ce0636" exitCode=0 Oct 03 17:22:45 crc kubenswrapper[4744]: I1003 17:22:45.833175 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vb729" event={"ID":"082b714a-fdd1-4c11-9759-1c826bbab583","Type":"ContainerDied","Data":"2e878649e63a60755f8f3b49a1a3c4c3989bf77069d87ca1080f58d551ce0636"} Oct 03 17:22:46 crc kubenswrapper[4744]: I1003 17:22:46.844431 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vb729" event={"ID":"082b714a-fdd1-4c11-9759-1c826bbab583","Type":"ContainerStarted","Data":"652fdd1eb1906342391fec7a585c92459b90642a90acd4060c164b7533ef206b"} Oct 03 17:22:46 crc kubenswrapper[4744]: I1003 17:22:46.865855 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vb729" podStartSLOduration=2.466841489 podStartE2EDuration="4.865837861s" podCreationTimestamp="2025-10-03 17:22:42 +0000 UTC" firstStartedPulling="2025-10-03 17:22:43.810115959 +0000 UTC m=+3490.089991855" lastFinishedPulling="2025-10-03 17:22:46.209112331 +0000 UTC m=+3492.488988227" observedRunningTime="2025-10-03 17:22:46.863450942 +0000 UTC m=+3493.143326838" watchObservedRunningTime="2025-10-03 17:22:46.865837861 +0000 UTC m=+3493.145713757" Oct 03 17:22:52 crc kubenswrapper[4744]: I1003 17:22:52.813129 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:52 crc kubenswrapper[4744]: I1003 17:22:52.816834 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:52 crc kubenswrapper[4744]: I1003 17:22:52.876710 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:52 crc kubenswrapper[4744]: I1003 17:22:52.944738 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:53 crc kubenswrapper[4744]: I1003 17:22:53.125922 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vb729"] Oct 03 17:22:54 crc kubenswrapper[4744]: I1003 17:22:54.911225 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vb729" podUID="082b714a-fdd1-4c11-9759-1c826bbab583" containerName="registry-server" containerID="cri-o://652fdd1eb1906342391fec7a585c92459b90642a90acd4060c164b7533ef206b" gracePeriod=2 Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.655787 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.836704 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdzxt\" (UniqueName: \"kubernetes.io/projected/082b714a-fdd1-4c11-9759-1c826bbab583-kube-api-access-pdzxt\") pod \"082b714a-fdd1-4c11-9759-1c826bbab583\" (UID: \"082b714a-fdd1-4c11-9759-1c826bbab583\") " Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.837062 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/082b714a-fdd1-4c11-9759-1c826bbab583-utilities\") pod \"082b714a-fdd1-4c11-9759-1c826bbab583\" (UID: \"082b714a-fdd1-4c11-9759-1c826bbab583\") " Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.837152 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/082b714a-fdd1-4c11-9759-1c826bbab583-catalog-content\") pod \"082b714a-fdd1-4c11-9759-1c826bbab583\" (UID: \"082b714a-fdd1-4c11-9759-1c826bbab583\") " Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.837828 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/082b714a-fdd1-4c11-9759-1c826bbab583-utilities" (OuterVolumeSpecName: "utilities") pod "082b714a-fdd1-4c11-9759-1c826bbab583" (UID: "082b714a-fdd1-4c11-9759-1c826bbab583"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.842213 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/082b714a-fdd1-4c11-9759-1c826bbab583-kube-api-access-pdzxt" (OuterVolumeSpecName: "kube-api-access-pdzxt") pod "082b714a-fdd1-4c11-9759-1c826bbab583" (UID: "082b714a-fdd1-4c11-9759-1c826bbab583"). InnerVolumeSpecName "kube-api-access-pdzxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.850063 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/082b714a-fdd1-4c11-9759-1c826bbab583-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "082b714a-fdd1-4c11-9759-1c826bbab583" (UID: "082b714a-fdd1-4c11-9759-1c826bbab583"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.920475 4744 generic.go:334] "Generic (PLEG): container finished" podID="082b714a-fdd1-4c11-9759-1c826bbab583" containerID="652fdd1eb1906342391fec7a585c92459b90642a90acd4060c164b7533ef206b" exitCode=0 Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.920522 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vb729" event={"ID":"082b714a-fdd1-4c11-9759-1c826bbab583","Type":"ContainerDied","Data":"652fdd1eb1906342391fec7a585c92459b90642a90acd4060c164b7533ef206b"} Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.920570 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vb729" event={"ID":"082b714a-fdd1-4c11-9759-1c826bbab583","Type":"ContainerDied","Data":"77291cc5cc35327a1420cc99d2430e3b3e869246cff6f94a2c21c1d8a44049e8"} Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.920572 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vb729" Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.920589 4744 scope.go:117] "RemoveContainer" containerID="652fdd1eb1906342391fec7a585c92459b90642a90acd4060c164b7533ef206b" Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.939713 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/082b714a-fdd1-4c11-9759-1c826bbab583-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.939739 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/082b714a-fdd1-4c11-9759-1c826bbab583-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.939751 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdzxt\" (UniqueName: \"kubernetes.io/projected/082b714a-fdd1-4c11-9759-1c826bbab583-kube-api-access-pdzxt\") on node \"crc\" DevicePath \"\"" Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.940724 4744 scope.go:117] "RemoveContainer" containerID="2e878649e63a60755f8f3b49a1a3c4c3989bf77069d87ca1080f58d551ce0636" Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.966583 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vb729"] Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.978239 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vb729"] Oct 03 17:22:55 crc kubenswrapper[4744]: I1003 17:22:55.979633 4744 scope.go:117] "RemoveContainer" containerID="5dce479679a12a2b14a8e82cfd87936a8710f94392af5d67734c36fa4542c285" Oct 03 17:22:56 crc kubenswrapper[4744]: I1003 17:22:56.004041 4744 scope.go:117] "RemoveContainer" containerID="652fdd1eb1906342391fec7a585c92459b90642a90acd4060c164b7533ef206b" Oct 03 17:22:56 crc kubenswrapper[4744]: E1003 17:22:56.004622 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"652fdd1eb1906342391fec7a585c92459b90642a90acd4060c164b7533ef206b\": container with ID starting with 652fdd1eb1906342391fec7a585c92459b90642a90acd4060c164b7533ef206b not found: ID does not exist" containerID="652fdd1eb1906342391fec7a585c92459b90642a90acd4060c164b7533ef206b" Oct 03 17:22:56 crc kubenswrapper[4744]: I1003 17:22:56.004758 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"652fdd1eb1906342391fec7a585c92459b90642a90acd4060c164b7533ef206b"} err="failed to get container status \"652fdd1eb1906342391fec7a585c92459b90642a90acd4060c164b7533ef206b\": rpc error: code = NotFound desc = could not find container \"652fdd1eb1906342391fec7a585c92459b90642a90acd4060c164b7533ef206b\": container with ID starting with 652fdd1eb1906342391fec7a585c92459b90642a90acd4060c164b7533ef206b not found: ID does not exist" Oct 03 17:22:56 crc kubenswrapper[4744]: I1003 17:22:56.004856 4744 scope.go:117] "RemoveContainer" containerID="2e878649e63a60755f8f3b49a1a3c4c3989bf77069d87ca1080f58d551ce0636" Oct 03 17:22:56 crc kubenswrapper[4744]: E1003 17:22:56.005825 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e878649e63a60755f8f3b49a1a3c4c3989bf77069d87ca1080f58d551ce0636\": container with ID starting with 2e878649e63a60755f8f3b49a1a3c4c3989bf77069d87ca1080f58d551ce0636 not found: ID does not exist" containerID="2e878649e63a60755f8f3b49a1a3c4c3989bf77069d87ca1080f58d551ce0636" Oct 03 17:22:56 crc kubenswrapper[4744]: I1003 17:22:56.005867 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e878649e63a60755f8f3b49a1a3c4c3989bf77069d87ca1080f58d551ce0636"} err="failed to get container status \"2e878649e63a60755f8f3b49a1a3c4c3989bf77069d87ca1080f58d551ce0636\": rpc error: code = NotFound desc = could not find container \"2e878649e63a60755f8f3b49a1a3c4c3989bf77069d87ca1080f58d551ce0636\": container with ID starting with 2e878649e63a60755f8f3b49a1a3c4c3989bf77069d87ca1080f58d551ce0636 not found: ID does not exist" Oct 03 17:22:56 crc kubenswrapper[4744]: I1003 17:22:56.005896 4744 scope.go:117] "RemoveContainer" containerID="5dce479679a12a2b14a8e82cfd87936a8710f94392af5d67734c36fa4542c285" Oct 03 17:22:56 crc kubenswrapper[4744]: E1003 17:22:56.006385 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dce479679a12a2b14a8e82cfd87936a8710f94392af5d67734c36fa4542c285\": container with ID starting with 5dce479679a12a2b14a8e82cfd87936a8710f94392af5d67734c36fa4542c285 not found: ID does not exist" containerID="5dce479679a12a2b14a8e82cfd87936a8710f94392af5d67734c36fa4542c285" Oct 03 17:22:56 crc kubenswrapper[4744]: I1003 17:22:56.006482 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dce479679a12a2b14a8e82cfd87936a8710f94392af5d67734c36fa4542c285"} err="failed to get container status \"5dce479679a12a2b14a8e82cfd87936a8710f94392af5d67734c36fa4542c285\": rpc error: code = NotFound desc = could not find container \"5dce479679a12a2b14a8e82cfd87936a8710f94392af5d67734c36fa4542c285\": container with ID starting with 5dce479679a12a2b14a8e82cfd87936a8710f94392af5d67734c36fa4542c285 not found: ID does not exist" Oct 03 17:22:56 crc kubenswrapper[4744]: I1003 17:22:56.911322 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="082b714a-fdd1-4c11-9759-1c826bbab583" path="/var/lib/kubelet/pods/082b714a-fdd1-4c11-9759-1c826bbab583/volumes" Oct 03 17:23:04 crc kubenswrapper[4744]: I1003 17:23:04.669767 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:23:04 crc kubenswrapper[4744]: I1003 17:23:04.670337 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:23:34 crc kubenswrapper[4744]: I1003 17:23:34.668871 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:23:34 crc kubenswrapper[4744]: I1003 17:23:34.669457 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:23:34 crc kubenswrapper[4744]: I1003 17:23:34.669538 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 17:23:34 crc kubenswrapper[4744]: I1003 17:23:34.670607 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cb4b2894f5fe1e71c4e0469f8b5747a451c56377648ebf3a4e30531cfe0daad2"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 17:23:34 crc kubenswrapper[4744]: I1003 17:23:34.670703 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://cb4b2894f5fe1e71c4e0469f8b5747a451c56377648ebf3a4e30531cfe0daad2" gracePeriod=600 Oct 03 17:23:35 crc kubenswrapper[4744]: I1003 17:23:35.270820 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="cb4b2894f5fe1e71c4e0469f8b5747a451c56377648ebf3a4e30531cfe0daad2" exitCode=0 Oct 03 17:23:35 crc kubenswrapper[4744]: I1003 17:23:35.270892 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"cb4b2894f5fe1e71c4e0469f8b5747a451c56377648ebf3a4e30531cfe0daad2"} Oct 03 17:23:35 crc kubenswrapper[4744]: I1003 17:23:35.271435 4744 scope.go:117] "RemoveContainer" containerID="6f7e5ecde9865ee8de4756773184d105cd97a86bb26ecc5b3b4525e3da007802" Oct 03 17:23:36 crc kubenswrapper[4744]: I1003 17:23:36.285577 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6"} Oct 03 17:26:04 crc kubenswrapper[4744]: I1003 17:26:04.668536 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:26:04 crc kubenswrapper[4744]: I1003 17:26:04.669081 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:26:34 crc kubenswrapper[4744]: I1003 17:26:34.668623 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:26:34 crc kubenswrapper[4744]: I1003 17:26:34.669155 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.775286 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lz2q4"] Oct 03 17:26:39 crc kubenswrapper[4744]: E1003 17:26:39.776410 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="082b714a-fdd1-4c11-9759-1c826bbab583" containerName="extract-utilities" Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.776424 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="082b714a-fdd1-4c11-9759-1c826bbab583" containerName="extract-utilities" Oct 03 17:26:39 crc kubenswrapper[4744]: E1003 17:26:39.776449 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="082b714a-fdd1-4c11-9759-1c826bbab583" containerName="registry-server" Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.776455 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="082b714a-fdd1-4c11-9759-1c826bbab583" containerName="registry-server" Oct 03 17:26:39 crc kubenswrapper[4744]: E1003 17:26:39.776486 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="082b714a-fdd1-4c11-9759-1c826bbab583" containerName="extract-content" Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.776511 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="082b714a-fdd1-4c11-9759-1c826bbab583" containerName="extract-content" Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.776721 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="082b714a-fdd1-4c11-9759-1c826bbab583" containerName="registry-server" Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.778461 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.785124 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lz2q4"] Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.834599 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ssmq\" (UniqueName: \"kubernetes.io/projected/f136f75e-a63f-4dce-8896-40a7e324a61b-kube-api-access-6ssmq\") pod \"certified-operators-lz2q4\" (UID: \"f136f75e-a63f-4dce-8896-40a7e324a61b\") " pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.834913 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f136f75e-a63f-4dce-8896-40a7e324a61b-catalog-content\") pod \"certified-operators-lz2q4\" (UID: \"f136f75e-a63f-4dce-8896-40a7e324a61b\") " pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.834961 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f136f75e-a63f-4dce-8896-40a7e324a61b-utilities\") pod \"certified-operators-lz2q4\" (UID: \"f136f75e-a63f-4dce-8896-40a7e324a61b\") " pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.936730 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ssmq\" (UniqueName: \"kubernetes.io/projected/f136f75e-a63f-4dce-8896-40a7e324a61b-kube-api-access-6ssmq\") pod \"certified-operators-lz2q4\" (UID: \"f136f75e-a63f-4dce-8896-40a7e324a61b\") " pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.936800 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f136f75e-a63f-4dce-8896-40a7e324a61b-catalog-content\") pod \"certified-operators-lz2q4\" (UID: \"f136f75e-a63f-4dce-8896-40a7e324a61b\") " pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.936842 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f136f75e-a63f-4dce-8896-40a7e324a61b-utilities\") pod \"certified-operators-lz2q4\" (UID: \"f136f75e-a63f-4dce-8896-40a7e324a61b\") " pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.937339 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f136f75e-a63f-4dce-8896-40a7e324a61b-catalog-content\") pod \"certified-operators-lz2q4\" (UID: \"f136f75e-a63f-4dce-8896-40a7e324a61b\") " pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.937372 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f136f75e-a63f-4dce-8896-40a7e324a61b-utilities\") pod \"certified-operators-lz2q4\" (UID: \"f136f75e-a63f-4dce-8896-40a7e324a61b\") " pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:39 crc kubenswrapper[4744]: I1003 17:26:39.973696 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ssmq\" (UniqueName: \"kubernetes.io/projected/f136f75e-a63f-4dce-8896-40a7e324a61b-kube-api-access-6ssmq\") pod \"certified-operators-lz2q4\" (UID: \"f136f75e-a63f-4dce-8896-40a7e324a61b\") " pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:40 crc kubenswrapper[4744]: I1003 17:26:40.102708 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:40 crc kubenswrapper[4744]: I1003 17:26:40.599677 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lz2q4"] Oct 03 17:26:41 crc kubenswrapper[4744]: I1003 17:26:41.100627 4744 generic.go:334] "Generic (PLEG): container finished" podID="f136f75e-a63f-4dce-8896-40a7e324a61b" containerID="8dfc24c45a65a0ba90f1f3a121d54b9c50c208d5313203319c762e8653902266" exitCode=0 Oct 03 17:26:41 crc kubenswrapper[4744]: I1003 17:26:41.100679 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz2q4" event={"ID":"f136f75e-a63f-4dce-8896-40a7e324a61b","Type":"ContainerDied","Data":"8dfc24c45a65a0ba90f1f3a121d54b9c50c208d5313203319c762e8653902266"} Oct 03 17:26:41 crc kubenswrapper[4744]: I1003 17:26:41.100709 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz2q4" event={"ID":"f136f75e-a63f-4dce-8896-40a7e324a61b","Type":"ContainerStarted","Data":"b34bc0162433d2f63d5edf27103434add9c4d920579f307f512303df0f3d5ad5"} Oct 03 17:26:41 crc kubenswrapper[4744]: I1003 17:26:41.102979 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 17:26:42 crc kubenswrapper[4744]: I1003 17:26:42.110513 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz2q4" event={"ID":"f136f75e-a63f-4dce-8896-40a7e324a61b","Type":"ContainerStarted","Data":"ba37ce4592fbacb7d387fd9bbef517f7b8c4cb8bc10e7957f9208a404f840b07"} Oct 03 17:26:44 crc kubenswrapper[4744]: I1003 17:26:44.134246 4744 generic.go:334] "Generic (PLEG): container finished" podID="f136f75e-a63f-4dce-8896-40a7e324a61b" containerID="ba37ce4592fbacb7d387fd9bbef517f7b8c4cb8bc10e7957f9208a404f840b07" exitCode=0 Oct 03 17:26:44 crc kubenswrapper[4744]: I1003 17:26:44.134356 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz2q4" event={"ID":"f136f75e-a63f-4dce-8896-40a7e324a61b","Type":"ContainerDied","Data":"ba37ce4592fbacb7d387fd9bbef517f7b8c4cb8bc10e7957f9208a404f840b07"} Oct 03 17:26:45 crc kubenswrapper[4744]: I1003 17:26:45.147888 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz2q4" event={"ID":"f136f75e-a63f-4dce-8896-40a7e324a61b","Type":"ContainerStarted","Data":"d7d6f71acb1528314531785d7031328a8d7cd0e083b7481f873a7d4b2e7754a1"} Oct 03 17:26:45 crc kubenswrapper[4744]: I1003 17:26:45.173625 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lz2q4" podStartSLOduration=2.6605777269999997 podStartE2EDuration="6.173601275s" podCreationTimestamp="2025-10-03 17:26:39 +0000 UTC" firstStartedPulling="2025-10-03 17:26:41.102705994 +0000 UTC m=+3727.382581900" lastFinishedPulling="2025-10-03 17:26:44.615729552 +0000 UTC m=+3730.895605448" observedRunningTime="2025-10-03 17:26:45.164831705 +0000 UTC m=+3731.444707621" watchObservedRunningTime="2025-10-03 17:26:45.173601275 +0000 UTC m=+3731.453477181" Oct 03 17:26:50 crc kubenswrapper[4744]: I1003 17:26:50.103905 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:50 crc kubenswrapper[4744]: I1003 17:26:50.104529 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:50 crc kubenswrapper[4744]: I1003 17:26:50.163521 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:50 crc kubenswrapper[4744]: I1003 17:26:50.251537 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:50 crc kubenswrapper[4744]: I1003 17:26:50.395536 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lz2q4"] Oct 03 17:26:52 crc kubenswrapper[4744]: I1003 17:26:52.220010 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lz2q4" podUID="f136f75e-a63f-4dce-8896-40a7e324a61b" containerName="registry-server" containerID="cri-o://d7d6f71acb1528314531785d7031328a8d7cd0e083b7481f873a7d4b2e7754a1" gracePeriod=2 Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.031812 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.201966 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f136f75e-a63f-4dce-8896-40a7e324a61b-utilities\") pod \"f136f75e-a63f-4dce-8896-40a7e324a61b\" (UID: \"f136f75e-a63f-4dce-8896-40a7e324a61b\") " Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.202050 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f136f75e-a63f-4dce-8896-40a7e324a61b-catalog-content\") pod \"f136f75e-a63f-4dce-8896-40a7e324a61b\" (UID: \"f136f75e-a63f-4dce-8896-40a7e324a61b\") " Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.202074 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ssmq\" (UniqueName: \"kubernetes.io/projected/f136f75e-a63f-4dce-8896-40a7e324a61b-kube-api-access-6ssmq\") pod \"f136f75e-a63f-4dce-8896-40a7e324a61b\" (UID: \"f136f75e-a63f-4dce-8896-40a7e324a61b\") " Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.203113 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f136f75e-a63f-4dce-8896-40a7e324a61b-utilities" (OuterVolumeSpecName: "utilities") pod "f136f75e-a63f-4dce-8896-40a7e324a61b" (UID: "f136f75e-a63f-4dce-8896-40a7e324a61b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.221953 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f136f75e-a63f-4dce-8896-40a7e324a61b-kube-api-access-6ssmq" (OuterVolumeSpecName: "kube-api-access-6ssmq") pod "f136f75e-a63f-4dce-8896-40a7e324a61b" (UID: "f136f75e-a63f-4dce-8896-40a7e324a61b"). InnerVolumeSpecName "kube-api-access-6ssmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.231056 4744 generic.go:334] "Generic (PLEG): container finished" podID="f136f75e-a63f-4dce-8896-40a7e324a61b" containerID="d7d6f71acb1528314531785d7031328a8d7cd0e083b7481f873a7d4b2e7754a1" exitCode=0 Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.231108 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz2q4" event={"ID":"f136f75e-a63f-4dce-8896-40a7e324a61b","Type":"ContainerDied","Data":"d7d6f71acb1528314531785d7031328a8d7cd0e083b7481f873a7d4b2e7754a1"} Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.231140 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lz2q4" event={"ID":"f136f75e-a63f-4dce-8896-40a7e324a61b","Type":"ContainerDied","Data":"b34bc0162433d2f63d5edf27103434add9c4d920579f307f512303df0f3d5ad5"} Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.231163 4744 scope.go:117] "RemoveContainer" containerID="d7d6f71acb1528314531785d7031328a8d7cd0e083b7481f873a7d4b2e7754a1" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.231208 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lz2q4" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.254882 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f136f75e-a63f-4dce-8896-40a7e324a61b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f136f75e-a63f-4dce-8896-40a7e324a61b" (UID: "f136f75e-a63f-4dce-8896-40a7e324a61b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.294471 4744 scope.go:117] "RemoveContainer" containerID="ba37ce4592fbacb7d387fd9bbef517f7b8c4cb8bc10e7957f9208a404f840b07" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.304099 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f136f75e-a63f-4dce-8896-40a7e324a61b-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.304148 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f136f75e-a63f-4dce-8896-40a7e324a61b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.304164 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ssmq\" (UniqueName: \"kubernetes.io/projected/f136f75e-a63f-4dce-8896-40a7e324a61b-kube-api-access-6ssmq\") on node \"crc\" DevicePath \"\"" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.321120 4744 scope.go:117] "RemoveContainer" containerID="8dfc24c45a65a0ba90f1f3a121d54b9c50c208d5313203319c762e8653902266" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.380666 4744 scope.go:117] "RemoveContainer" containerID="d7d6f71acb1528314531785d7031328a8d7cd0e083b7481f873a7d4b2e7754a1" Oct 03 17:26:53 crc kubenswrapper[4744]: E1003 17:26:53.381532 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7d6f71acb1528314531785d7031328a8d7cd0e083b7481f873a7d4b2e7754a1\": container with ID starting with d7d6f71acb1528314531785d7031328a8d7cd0e083b7481f873a7d4b2e7754a1 not found: ID does not exist" containerID="d7d6f71acb1528314531785d7031328a8d7cd0e083b7481f873a7d4b2e7754a1" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.381580 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7d6f71acb1528314531785d7031328a8d7cd0e083b7481f873a7d4b2e7754a1"} err="failed to get container status \"d7d6f71acb1528314531785d7031328a8d7cd0e083b7481f873a7d4b2e7754a1\": rpc error: code = NotFound desc = could not find container \"d7d6f71acb1528314531785d7031328a8d7cd0e083b7481f873a7d4b2e7754a1\": container with ID starting with d7d6f71acb1528314531785d7031328a8d7cd0e083b7481f873a7d4b2e7754a1 not found: ID does not exist" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.381614 4744 scope.go:117] "RemoveContainer" containerID="ba37ce4592fbacb7d387fd9bbef517f7b8c4cb8bc10e7957f9208a404f840b07" Oct 03 17:26:53 crc kubenswrapper[4744]: E1003 17:26:53.381940 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba37ce4592fbacb7d387fd9bbef517f7b8c4cb8bc10e7957f9208a404f840b07\": container with ID starting with ba37ce4592fbacb7d387fd9bbef517f7b8c4cb8bc10e7957f9208a404f840b07 not found: ID does not exist" containerID="ba37ce4592fbacb7d387fd9bbef517f7b8c4cb8bc10e7957f9208a404f840b07" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.381990 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba37ce4592fbacb7d387fd9bbef517f7b8c4cb8bc10e7957f9208a404f840b07"} err="failed to get container status \"ba37ce4592fbacb7d387fd9bbef517f7b8c4cb8bc10e7957f9208a404f840b07\": rpc error: code = NotFound desc = could not find container \"ba37ce4592fbacb7d387fd9bbef517f7b8c4cb8bc10e7957f9208a404f840b07\": container with ID starting with ba37ce4592fbacb7d387fd9bbef517f7b8c4cb8bc10e7957f9208a404f840b07 not found: ID does not exist" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.382024 4744 scope.go:117] "RemoveContainer" containerID="8dfc24c45a65a0ba90f1f3a121d54b9c50c208d5313203319c762e8653902266" Oct 03 17:26:53 crc kubenswrapper[4744]: E1003 17:26:53.382456 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dfc24c45a65a0ba90f1f3a121d54b9c50c208d5313203319c762e8653902266\": container with ID starting with 8dfc24c45a65a0ba90f1f3a121d54b9c50c208d5313203319c762e8653902266 not found: ID does not exist" containerID="8dfc24c45a65a0ba90f1f3a121d54b9c50c208d5313203319c762e8653902266" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.382489 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dfc24c45a65a0ba90f1f3a121d54b9c50c208d5313203319c762e8653902266"} err="failed to get container status \"8dfc24c45a65a0ba90f1f3a121d54b9c50c208d5313203319c762e8653902266\": rpc error: code = NotFound desc = could not find container \"8dfc24c45a65a0ba90f1f3a121d54b9c50c208d5313203319c762e8653902266\": container with ID starting with 8dfc24c45a65a0ba90f1f3a121d54b9c50c208d5313203319c762e8653902266 not found: ID does not exist" Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.563342 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lz2q4"] Oct 03 17:26:53 crc kubenswrapper[4744]: I1003 17:26:53.571993 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lz2q4"] Oct 03 17:26:54 crc kubenswrapper[4744]: I1003 17:26:54.908181 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f136f75e-a63f-4dce-8896-40a7e324a61b" path="/var/lib/kubelet/pods/f136f75e-a63f-4dce-8896-40a7e324a61b/volumes" Oct 03 17:27:04 crc kubenswrapper[4744]: I1003 17:27:04.668451 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:27:04 crc kubenswrapper[4744]: I1003 17:27:04.669047 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:27:04 crc kubenswrapper[4744]: I1003 17:27:04.669090 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 17:27:04 crc kubenswrapper[4744]: I1003 17:27:04.669975 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 17:27:04 crc kubenswrapper[4744]: I1003 17:27:04.670045 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" gracePeriod=600 Oct 03 17:27:04 crc kubenswrapper[4744]: E1003 17:27:04.811947 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:27:05 crc kubenswrapper[4744]: I1003 17:27:05.352819 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" exitCode=0 Oct 03 17:27:05 crc kubenswrapper[4744]: I1003 17:27:05.352866 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6"} Oct 03 17:27:05 crc kubenswrapper[4744]: I1003 17:27:05.353225 4744 scope.go:117] "RemoveContainer" containerID="cb4b2894f5fe1e71c4e0469f8b5747a451c56377648ebf3a4e30531cfe0daad2" Oct 03 17:27:05 crc kubenswrapper[4744]: I1003 17:27:05.354068 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:27:05 crc kubenswrapper[4744]: E1003 17:27:05.354406 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:27:17 crc kubenswrapper[4744]: I1003 17:27:17.891926 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:27:17 crc kubenswrapper[4744]: E1003 17:27:17.892707 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:27:19 crc kubenswrapper[4744]: I1003 17:27:19.503543 4744 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-4n9cl" Oct 03 17:27:32 crc kubenswrapper[4744]: I1003 17:27:32.891845 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:27:32 crc kubenswrapper[4744]: E1003 17:27:32.892708 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:27:46 crc kubenswrapper[4744]: I1003 17:27:46.892380 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:27:46 crc kubenswrapper[4744]: E1003 17:27:46.893356 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:27:59 crc kubenswrapper[4744]: I1003 17:27:59.893742 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:27:59 crc kubenswrapper[4744]: E1003 17:27:59.895039 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:28:14 crc kubenswrapper[4744]: I1003 17:28:14.898667 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:28:14 crc kubenswrapper[4744]: E1003 17:28:14.899585 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:28:27 crc kubenswrapper[4744]: I1003 17:28:27.892438 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:28:27 crc kubenswrapper[4744]: E1003 17:28:27.893352 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:28:42 crc kubenswrapper[4744]: I1003 17:28:42.892984 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:28:42 crc kubenswrapper[4744]: E1003 17:28:42.894419 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:28:54 crc kubenswrapper[4744]: I1003 17:28:54.898225 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:28:54 crc kubenswrapper[4744]: E1003 17:28:54.900012 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:29:09 crc kubenswrapper[4744]: I1003 17:29:09.892093 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:29:09 crc kubenswrapper[4744]: E1003 17:29:09.893168 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:29:22 crc kubenswrapper[4744]: I1003 17:29:22.892148 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:29:22 crc kubenswrapper[4744]: E1003 17:29:22.893136 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:29:29 crc kubenswrapper[4744]: I1003 17:29:29.830773 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5pp8j"] Oct 03 17:29:29 crc kubenswrapper[4744]: E1003 17:29:29.831475 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f136f75e-a63f-4dce-8896-40a7e324a61b" containerName="extract-utilities" Oct 03 17:29:29 crc kubenswrapper[4744]: I1003 17:29:29.831486 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f136f75e-a63f-4dce-8896-40a7e324a61b" containerName="extract-utilities" Oct 03 17:29:29 crc kubenswrapper[4744]: E1003 17:29:29.831510 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f136f75e-a63f-4dce-8896-40a7e324a61b" containerName="extract-content" Oct 03 17:29:29 crc kubenswrapper[4744]: I1003 17:29:29.831516 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f136f75e-a63f-4dce-8896-40a7e324a61b" containerName="extract-content" Oct 03 17:29:29 crc kubenswrapper[4744]: E1003 17:29:29.831538 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f136f75e-a63f-4dce-8896-40a7e324a61b" containerName="registry-server" Oct 03 17:29:29 crc kubenswrapper[4744]: I1003 17:29:29.831544 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f136f75e-a63f-4dce-8896-40a7e324a61b" containerName="registry-server" Oct 03 17:29:29 crc kubenswrapper[4744]: I1003 17:29:29.831727 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f136f75e-a63f-4dce-8896-40a7e324a61b" containerName="registry-server" Oct 03 17:29:29 crc kubenswrapper[4744]: I1003 17:29:29.832993 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:29 crc kubenswrapper[4744]: I1003 17:29:29.843167 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5pp8j"] Oct 03 17:29:29 crc kubenswrapper[4744]: I1003 17:29:29.913937 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr6zh\" (UniqueName: \"kubernetes.io/projected/da363856-90eb-4f56-9247-5ea26fe270fc-kube-api-access-hr6zh\") pod \"community-operators-5pp8j\" (UID: \"da363856-90eb-4f56-9247-5ea26fe270fc\") " pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:29 crc kubenswrapper[4744]: I1003 17:29:29.914072 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da363856-90eb-4f56-9247-5ea26fe270fc-utilities\") pod \"community-operators-5pp8j\" (UID: \"da363856-90eb-4f56-9247-5ea26fe270fc\") " pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:29 crc kubenswrapper[4744]: I1003 17:29:29.914793 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da363856-90eb-4f56-9247-5ea26fe270fc-catalog-content\") pod \"community-operators-5pp8j\" (UID: \"da363856-90eb-4f56-9247-5ea26fe270fc\") " pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:30 crc kubenswrapper[4744]: I1003 17:29:30.016391 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da363856-90eb-4f56-9247-5ea26fe270fc-catalog-content\") pod \"community-operators-5pp8j\" (UID: \"da363856-90eb-4f56-9247-5ea26fe270fc\") " pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:30 crc kubenswrapper[4744]: I1003 17:29:30.016508 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr6zh\" (UniqueName: \"kubernetes.io/projected/da363856-90eb-4f56-9247-5ea26fe270fc-kube-api-access-hr6zh\") pod \"community-operators-5pp8j\" (UID: \"da363856-90eb-4f56-9247-5ea26fe270fc\") " pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:30 crc kubenswrapper[4744]: I1003 17:29:30.016581 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da363856-90eb-4f56-9247-5ea26fe270fc-utilities\") pod \"community-operators-5pp8j\" (UID: \"da363856-90eb-4f56-9247-5ea26fe270fc\") " pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:30 crc kubenswrapper[4744]: I1003 17:29:30.016919 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da363856-90eb-4f56-9247-5ea26fe270fc-catalog-content\") pod \"community-operators-5pp8j\" (UID: \"da363856-90eb-4f56-9247-5ea26fe270fc\") " pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:30 crc kubenswrapper[4744]: I1003 17:29:30.016997 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da363856-90eb-4f56-9247-5ea26fe270fc-utilities\") pod \"community-operators-5pp8j\" (UID: \"da363856-90eb-4f56-9247-5ea26fe270fc\") " pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:30 crc kubenswrapper[4744]: I1003 17:29:30.045880 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr6zh\" (UniqueName: \"kubernetes.io/projected/da363856-90eb-4f56-9247-5ea26fe270fc-kube-api-access-hr6zh\") pod \"community-operators-5pp8j\" (UID: \"da363856-90eb-4f56-9247-5ea26fe270fc\") " pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:30 crc kubenswrapper[4744]: I1003 17:29:30.177105 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:30 crc kubenswrapper[4744]: I1003 17:29:30.951209 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5pp8j"] Oct 03 17:29:31 crc kubenswrapper[4744]: I1003 17:29:31.788615 4744 generic.go:334] "Generic (PLEG): container finished" podID="da363856-90eb-4f56-9247-5ea26fe270fc" containerID="62ba469fd355a97157fb835d24d474847b7bee94c5e9db48362dbcd4ee4659f4" exitCode=0 Oct 03 17:29:31 crc kubenswrapper[4744]: I1003 17:29:31.788808 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pp8j" event={"ID":"da363856-90eb-4f56-9247-5ea26fe270fc","Type":"ContainerDied","Data":"62ba469fd355a97157fb835d24d474847b7bee94c5e9db48362dbcd4ee4659f4"} Oct 03 17:29:31 crc kubenswrapper[4744]: I1003 17:29:31.789166 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pp8j" event={"ID":"da363856-90eb-4f56-9247-5ea26fe270fc","Type":"ContainerStarted","Data":"0cfef19ec9b7464fbe171ea2aa4fc1bca7ba95c7b093af873b52711fe896ef64"} Oct 03 17:29:32 crc kubenswrapper[4744]: I1003 17:29:32.799634 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pp8j" event={"ID":"da363856-90eb-4f56-9247-5ea26fe270fc","Type":"ContainerStarted","Data":"37933b09a3c29910e2a6c3a1bbfc8b4edd2b21968cbff7652023f684c21936c6"} Oct 03 17:29:34 crc kubenswrapper[4744]: I1003 17:29:34.819303 4744 generic.go:334] "Generic (PLEG): container finished" podID="da363856-90eb-4f56-9247-5ea26fe270fc" containerID="37933b09a3c29910e2a6c3a1bbfc8b4edd2b21968cbff7652023f684c21936c6" exitCode=0 Oct 03 17:29:34 crc kubenswrapper[4744]: I1003 17:29:34.819405 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pp8j" event={"ID":"da363856-90eb-4f56-9247-5ea26fe270fc","Type":"ContainerDied","Data":"37933b09a3c29910e2a6c3a1bbfc8b4edd2b21968cbff7652023f684c21936c6"} Oct 03 17:29:34 crc kubenswrapper[4744]: I1003 17:29:34.934644 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:29:34 crc kubenswrapper[4744]: E1003 17:29:34.935452 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:29:35 crc kubenswrapper[4744]: I1003 17:29:35.829918 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pp8j" event={"ID":"da363856-90eb-4f56-9247-5ea26fe270fc","Type":"ContainerStarted","Data":"0f316b879cf41e3974eef4415e69bb6f0ebd16f562219c23bf27983f65317cf5"} Oct 03 17:29:35 crc kubenswrapper[4744]: I1003 17:29:35.849659 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5pp8j" podStartSLOduration=3.440510434 podStartE2EDuration="6.849637629s" podCreationTimestamp="2025-10-03 17:29:29 +0000 UTC" firstStartedPulling="2025-10-03 17:29:31.792276406 +0000 UTC m=+3898.072152292" lastFinishedPulling="2025-10-03 17:29:35.201403591 +0000 UTC m=+3901.481279487" observedRunningTime="2025-10-03 17:29:35.848329376 +0000 UTC m=+3902.128205272" watchObservedRunningTime="2025-10-03 17:29:35.849637629 +0000 UTC m=+3902.129513525" Oct 03 17:29:40 crc kubenswrapper[4744]: I1003 17:29:40.178127 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:40 crc kubenswrapper[4744]: I1003 17:29:40.178668 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:40 crc kubenswrapper[4744]: I1003 17:29:40.227677 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:40 crc kubenswrapper[4744]: I1003 17:29:40.940224 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:41 crc kubenswrapper[4744]: I1003 17:29:41.008439 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5pp8j"] Oct 03 17:29:42 crc kubenswrapper[4744]: I1003 17:29:42.896153 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5pp8j" podUID="da363856-90eb-4f56-9247-5ea26fe270fc" containerName="registry-server" containerID="cri-o://0f316b879cf41e3974eef4415e69bb6f0ebd16f562219c23bf27983f65317cf5" gracePeriod=2 Oct 03 17:29:43 crc kubenswrapper[4744]: I1003 17:29:43.627486 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:43 crc kubenswrapper[4744]: I1003 17:29:43.798206 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da363856-90eb-4f56-9247-5ea26fe270fc-utilities\") pod \"da363856-90eb-4f56-9247-5ea26fe270fc\" (UID: \"da363856-90eb-4f56-9247-5ea26fe270fc\") " Oct 03 17:29:43 crc kubenswrapper[4744]: I1003 17:29:43.798297 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr6zh\" (UniqueName: \"kubernetes.io/projected/da363856-90eb-4f56-9247-5ea26fe270fc-kube-api-access-hr6zh\") pod \"da363856-90eb-4f56-9247-5ea26fe270fc\" (UID: \"da363856-90eb-4f56-9247-5ea26fe270fc\") " Oct 03 17:29:43 crc kubenswrapper[4744]: I1003 17:29:43.798384 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da363856-90eb-4f56-9247-5ea26fe270fc-catalog-content\") pod \"da363856-90eb-4f56-9247-5ea26fe270fc\" (UID: \"da363856-90eb-4f56-9247-5ea26fe270fc\") " Oct 03 17:29:43 crc kubenswrapper[4744]: I1003 17:29:43.799381 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da363856-90eb-4f56-9247-5ea26fe270fc-utilities" (OuterVolumeSpecName: "utilities") pod "da363856-90eb-4f56-9247-5ea26fe270fc" (UID: "da363856-90eb-4f56-9247-5ea26fe270fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:29:43 crc kubenswrapper[4744]: I1003 17:29:43.814986 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da363856-90eb-4f56-9247-5ea26fe270fc-kube-api-access-hr6zh" (OuterVolumeSpecName: "kube-api-access-hr6zh") pod "da363856-90eb-4f56-9247-5ea26fe270fc" (UID: "da363856-90eb-4f56-9247-5ea26fe270fc"). InnerVolumeSpecName "kube-api-access-hr6zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:29:43 crc kubenswrapper[4744]: I1003 17:29:43.902577 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da363856-90eb-4f56-9247-5ea26fe270fc-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:29:43 crc kubenswrapper[4744]: I1003 17:29:43.902629 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr6zh\" (UniqueName: \"kubernetes.io/projected/da363856-90eb-4f56-9247-5ea26fe270fc-kube-api-access-hr6zh\") on node \"crc\" DevicePath \"\"" Oct 03 17:29:43 crc kubenswrapper[4744]: I1003 17:29:43.908009 4744 generic.go:334] "Generic (PLEG): container finished" podID="da363856-90eb-4f56-9247-5ea26fe270fc" containerID="0f316b879cf41e3974eef4415e69bb6f0ebd16f562219c23bf27983f65317cf5" exitCode=0 Oct 03 17:29:43 crc kubenswrapper[4744]: I1003 17:29:43.908070 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pp8j" event={"ID":"da363856-90eb-4f56-9247-5ea26fe270fc","Type":"ContainerDied","Data":"0f316b879cf41e3974eef4415e69bb6f0ebd16f562219c23bf27983f65317cf5"} Oct 03 17:29:43 crc kubenswrapper[4744]: I1003 17:29:43.908121 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5pp8j" Oct 03 17:29:43 crc kubenswrapper[4744]: I1003 17:29:43.908167 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pp8j" event={"ID":"da363856-90eb-4f56-9247-5ea26fe270fc","Type":"ContainerDied","Data":"0cfef19ec9b7464fbe171ea2aa4fc1bca7ba95c7b093af873b52711fe896ef64"} Oct 03 17:29:43 crc kubenswrapper[4744]: I1003 17:29:43.908203 4744 scope.go:117] "RemoveContainer" containerID="0f316b879cf41e3974eef4415e69bb6f0ebd16f562219c23bf27983f65317cf5" Oct 03 17:29:43 crc kubenswrapper[4744]: I1003 17:29:43.928843 4744 scope.go:117] "RemoveContainer" containerID="37933b09a3c29910e2a6c3a1bbfc8b4edd2b21968cbff7652023f684c21936c6" Oct 03 17:29:43 crc kubenswrapper[4744]: I1003 17:29:43.961428 4744 scope.go:117] "RemoveContainer" containerID="62ba469fd355a97157fb835d24d474847b7bee94c5e9db48362dbcd4ee4659f4" Oct 03 17:29:44 crc kubenswrapper[4744]: I1003 17:29:44.010979 4744 scope.go:117] "RemoveContainer" containerID="0f316b879cf41e3974eef4415e69bb6f0ebd16f562219c23bf27983f65317cf5" Oct 03 17:29:44 crc kubenswrapper[4744]: E1003 17:29:44.011455 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f316b879cf41e3974eef4415e69bb6f0ebd16f562219c23bf27983f65317cf5\": container with ID starting with 0f316b879cf41e3974eef4415e69bb6f0ebd16f562219c23bf27983f65317cf5 not found: ID does not exist" containerID="0f316b879cf41e3974eef4415e69bb6f0ebd16f562219c23bf27983f65317cf5" Oct 03 17:29:44 crc kubenswrapper[4744]: I1003 17:29:44.011550 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f316b879cf41e3974eef4415e69bb6f0ebd16f562219c23bf27983f65317cf5"} err="failed to get container status \"0f316b879cf41e3974eef4415e69bb6f0ebd16f562219c23bf27983f65317cf5\": rpc error: code = NotFound desc = could not find container \"0f316b879cf41e3974eef4415e69bb6f0ebd16f562219c23bf27983f65317cf5\": container with ID starting with 0f316b879cf41e3974eef4415e69bb6f0ebd16f562219c23bf27983f65317cf5 not found: ID does not exist" Oct 03 17:29:44 crc kubenswrapper[4744]: I1003 17:29:44.011629 4744 scope.go:117] "RemoveContainer" containerID="37933b09a3c29910e2a6c3a1bbfc8b4edd2b21968cbff7652023f684c21936c6" Oct 03 17:29:44 crc kubenswrapper[4744]: E1003 17:29:44.012147 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37933b09a3c29910e2a6c3a1bbfc8b4edd2b21968cbff7652023f684c21936c6\": container with ID starting with 37933b09a3c29910e2a6c3a1bbfc8b4edd2b21968cbff7652023f684c21936c6 not found: ID does not exist" containerID="37933b09a3c29910e2a6c3a1bbfc8b4edd2b21968cbff7652023f684c21936c6" Oct 03 17:29:44 crc kubenswrapper[4744]: I1003 17:29:44.012195 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37933b09a3c29910e2a6c3a1bbfc8b4edd2b21968cbff7652023f684c21936c6"} err="failed to get container status \"37933b09a3c29910e2a6c3a1bbfc8b4edd2b21968cbff7652023f684c21936c6\": rpc error: code = NotFound desc = could not find container \"37933b09a3c29910e2a6c3a1bbfc8b4edd2b21968cbff7652023f684c21936c6\": container with ID starting with 37933b09a3c29910e2a6c3a1bbfc8b4edd2b21968cbff7652023f684c21936c6 not found: ID does not exist" Oct 03 17:29:44 crc kubenswrapper[4744]: I1003 17:29:44.012225 4744 scope.go:117] "RemoveContainer" containerID="62ba469fd355a97157fb835d24d474847b7bee94c5e9db48362dbcd4ee4659f4" Oct 03 17:29:44 crc kubenswrapper[4744]: E1003 17:29:44.012518 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62ba469fd355a97157fb835d24d474847b7bee94c5e9db48362dbcd4ee4659f4\": container with ID starting with 62ba469fd355a97157fb835d24d474847b7bee94c5e9db48362dbcd4ee4659f4 not found: ID does not exist" containerID="62ba469fd355a97157fb835d24d474847b7bee94c5e9db48362dbcd4ee4659f4" Oct 03 17:29:44 crc kubenswrapper[4744]: I1003 17:29:44.012548 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62ba469fd355a97157fb835d24d474847b7bee94c5e9db48362dbcd4ee4659f4"} err="failed to get container status \"62ba469fd355a97157fb835d24d474847b7bee94c5e9db48362dbcd4ee4659f4\": rpc error: code = NotFound desc = could not find container \"62ba469fd355a97157fb835d24d474847b7bee94c5e9db48362dbcd4ee4659f4\": container with ID starting with 62ba469fd355a97157fb835d24d474847b7bee94c5e9db48362dbcd4ee4659f4 not found: ID does not exist" Oct 03 17:29:44 crc kubenswrapper[4744]: I1003 17:29:44.205244 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da363856-90eb-4f56-9247-5ea26fe270fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da363856-90eb-4f56-9247-5ea26fe270fc" (UID: "da363856-90eb-4f56-9247-5ea26fe270fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:29:44 crc kubenswrapper[4744]: I1003 17:29:44.209524 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da363856-90eb-4f56-9247-5ea26fe270fc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:29:44 crc kubenswrapper[4744]: I1003 17:29:44.252285 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5pp8j"] Oct 03 17:29:44 crc kubenswrapper[4744]: I1003 17:29:44.262822 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5pp8j"] Oct 03 17:29:44 crc kubenswrapper[4744]: I1003 17:29:44.904339 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da363856-90eb-4f56-9247-5ea26fe270fc" path="/var/lib/kubelet/pods/da363856-90eb-4f56-9247-5ea26fe270fc/volumes" Oct 03 17:29:45 crc kubenswrapper[4744]: I1003 17:29:45.892420 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:29:45 crc kubenswrapper[4744]: E1003 17:29:45.892925 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:29:56 crc kubenswrapper[4744]: I1003 17:29:56.891483 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:29:56 crc kubenswrapper[4744]: E1003 17:29:56.892241 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.147627 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs"] Oct 03 17:30:00 crc kubenswrapper[4744]: E1003 17:30:00.148750 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da363856-90eb-4f56-9247-5ea26fe270fc" containerName="extract-content" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.148768 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="da363856-90eb-4f56-9247-5ea26fe270fc" containerName="extract-content" Oct 03 17:30:00 crc kubenswrapper[4744]: E1003 17:30:00.148826 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da363856-90eb-4f56-9247-5ea26fe270fc" containerName="registry-server" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.148835 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="da363856-90eb-4f56-9247-5ea26fe270fc" containerName="registry-server" Oct 03 17:30:00 crc kubenswrapper[4744]: E1003 17:30:00.148849 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da363856-90eb-4f56-9247-5ea26fe270fc" containerName="extract-utilities" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.148857 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="da363856-90eb-4f56-9247-5ea26fe270fc" containerName="extract-utilities" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.149120 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="da363856-90eb-4f56-9247-5ea26fe270fc" containerName="registry-server" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.150020 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.152398 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.154869 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.177420 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs"] Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.244609 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e978d1e-b25a-4c13-a138-4bdc309ca419-secret-volume\") pod \"collect-profiles-29325210-6q8xs\" (UID: \"5e978d1e-b25a-4c13-a138-4bdc309ca419\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.244758 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e978d1e-b25a-4c13-a138-4bdc309ca419-config-volume\") pod \"collect-profiles-29325210-6q8xs\" (UID: \"5e978d1e-b25a-4c13-a138-4bdc309ca419\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.244800 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46xs5\" (UniqueName: \"kubernetes.io/projected/5e978d1e-b25a-4c13-a138-4bdc309ca419-kube-api-access-46xs5\") pod \"collect-profiles-29325210-6q8xs\" (UID: \"5e978d1e-b25a-4c13-a138-4bdc309ca419\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.353954 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e978d1e-b25a-4c13-a138-4bdc309ca419-secret-volume\") pod \"collect-profiles-29325210-6q8xs\" (UID: \"5e978d1e-b25a-4c13-a138-4bdc309ca419\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.354129 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e978d1e-b25a-4c13-a138-4bdc309ca419-config-volume\") pod \"collect-profiles-29325210-6q8xs\" (UID: \"5e978d1e-b25a-4c13-a138-4bdc309ca419\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.354189 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46xs5\" (UniqueName: \"kubernetes.io/projected/5e978d1e-b25a-4c13-a138-4bdc309ca419-kube-api-access-46xs5\") pod \"collect-profiles-29325210-6q8xs\" (UID: \"5e978d1e-b25a-4c13-a138-4bdc309ca419\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.355019 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e978d1e-b25a-4c13-a138-4bdc309ca419-config-volume\") pod \"collect-profiles-29325210-6q8xs\" (UID: \"5e978d1e-b25a-4c13-a138-4bdc309ca419\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.368105 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e978d1e-b25a-4c13-a138-4bdc309ca419-secret-volume\") pod \"collect-profiles-29325210-6q8xs\" (UID: \"5e978d1e-b25a-4c13-a138-4bdc309ca419\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.382531 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46xs5\" (UniqueName: \"kubernetes.io/projected/5e978d1e-b25a-4c13-a138-4bdc309ca419-kube-api-access-46xs5\") pod \"collect-profiles-29325210-6q8xs\" (UID: \"5e978d1e-b25a-4c13-a138-4bdc309ca419\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" Oct 03 17:30:00 crc kubenswrapper[4744]: I1003 17:30:00.489222 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" Oct 03 17:30:01 crc kubenswrapper[4744]: I1003 17:30:01.005864 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs"] Oct 03 17:30:01 crc kubenswrapper[4744]: I1003 17:30:01.086380 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" event={"ID":"5e978d1e-b25a-4c13-a138-4bdc309ca419","Type":"ContainerStarted","Data":"b768a99f68dc6841c5140548ea826e46a28662c00a40b38592db4b0042e374b7"} Oct 03 17:30:02 crc kubenswrapper[4744]: I1003 17:30:02.097354 4744 generic.go:334] "Generic (PLEG): container finished" podID="5e978d1e-b25a-4c13-a138-4bdc309ca419" containerID="54874c6e9bee1d196bd8f9698b9afa5c5148ffad161af4e29a7ebd46aa61b4e5" exitCode=0 Oct 03 17:30:02 crc kubenswrapper[4744]: I1003 17:30:02.097443 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" event={"ID":"5e978d1e-b25a-4c13-a138-4bdc309ca419","Type":"ContainerDied","Data":"54874c6e9bee1d196bd8f9698b9afa5c5148ffad161af4e29a7ebd46aa61b4e5"} Oct 03 17:30:03 crc kubenswrapper[4744]: I1003 17:30:03.773102 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" Oct 03 17:30:03 crc kubenswrapper[4744]: I1003 17:30:03.824876 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e978d1e-b25a-4c13-a138-4bdc309ca419-config-volume\") pod \"5e978d1e-b25a-4c13-a138-4bdc309ca419\" (UID: \"5e978d1e-b25a-4c13-a138-4bdc309ca419\") " Oct 03 17:30:03 crc kubenswrapper[4744]: I1003 17:30:03.824973 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e978d1e-b25a-4c13-a138-4bdc309ca419-secret-volume\") pod \"5e978d1e-b25a-4c13-a138-4bdc309ca419\" (UID: \"5e978d1e-b25a-4c13-a138-4bdc309ca419\") " Oct 03 17:30:03 crc kubenswrapper[4744]: I1003 17:30:03.825109 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46xs5\" (UniqueName: \"kubernetes.io/projected/5e978d1e-b25a-4c13-a138-4bdc309ca419-kube-api-access-46xs5\") pod \"5e978d1e-b25a-4c13-a138-4bdc309ca419\" (UID: \"5e978d1e-b25a-4c13-a138-4bdc309ca419\") " Oct 03 17:30:03 crc kubenswrapper[4744]: I1003 17:30:03.825789 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e978d1e-b25a-4c13-a138-4bdc309ca419-config-volume" (OuterVolumeSpecName: "config-volume") pod "5e978d1e-b25a-4c13-a138-4bdc309ca419" (UID: "5e978d1e-b25a-4c13-a138-4bdc309ca419"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 17:30:03 crc kubenswrapper[4744]: I1003 17:30:03.831409 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e978d1e-b25a-4c13-a138-4bdc309ca419-kube-api-access-46xs5" (OuterVolumeSpecName: "kube-api-access-46xs5") pod "5e978d1e-b25a-4c13-a138-4bdc309ca419" (UID: "5e978d1e-b25a-4c13-a138-4bdc309ca419"). InnerVolumeSpecName "kube-api-access-46xs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:30:03 crc kubenswrapper[4744]: I1003 17:30:03.838725 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e978d1e-b25a-4c13-a138-4bdc309ca419-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5e978d1e-b25a-4c13-a138-4bdc309ca419" (UID: "5e978d1e-b25a-4c13-a138-4bdc309ca419"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:30:03 crc kubenswrapper[4744]: I1003 17:30:03.927801 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46xs5\" (UniqueName: \"kubernetes.io/projected/5e978d1e-b25a-4c13-a138-4bdc309ca419-kube-api-access-46xs5\") on node \"crc\" DevicePath \"\"" Oct 03 17:30:03 crc kubenswrapper[4744]: I1003 17:30:03.927840 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e978d1e-b25a-4c13-a138-4bdc309ca419-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 17:30:03 crc kubenswrapper[4744]: I1003 17:30:03.927852 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e978d1e-b25a-4c13-a138-4bdc309ca419-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 17:30:04 crc kubenswrapper[4744]: I1003 17:30:04.116238 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" event={"ID":"5e978d1e-b25a-4c13-a138-4bdc309ca419","Type":"ContainerDied","Data":"b768a99f68dc6841c5140548ea826e46a28662c00a40b38592db4b0042e374b7"} Oct 03 17:30:04 crc kubenswrapper[4744]: I1003 17:30:04.116271 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b768a99f68dc6841c5140548ea826e46a28662c00a40b38592db4b0042e374b7" Oct 03 17:30:04 crc kubenswrapper[4744]: I1003 17:30:04.116320 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325210-6q8xs" Oct 03 17:30:04 crc kubenswrapper[4744]: I1003 17:30:04.842432 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr"] Oct 03 17:30:04 crc kubenswrapper[4744]: I1003 17:30:04.850459 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325165-bxpsr"] Oct 03 17:30:04 crc kubenswrapper[4744]: I1003 17:30:04.907670 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="451ef308-a0a4-4cf2-9132-7a6c40a9b35c" path="/var/lib/kubelet/pods/451ef308-a0a4-4cf2-9132-7a6c40a9b35c/volumes" Oct 03 17:30:11 crc kubenswrapper[4744]: I1003 17:30:11.891718 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:30:11 crc kubenswrapper[4744]: E1003 17:30:11.892615 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:30:26 crc kubenswrapper[4744]: I1003 17:30:26.891931 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:30:26 crc kubenswrapper[4744]: E1003 17:30:26.892825 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:30:38 crc kubenswrapper[4744]: I1003 17:30:38.892685 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:30:38 crc kubenswrapper[4744]: E1003 17:30:38.894969 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:30:40 crc kubenswrapper[4744]: I1003 17:30:40.699908 4744 scope.go:117] "RemoveContainer" containerID="c7fdb9ddea7fb74ac8e2983a1b1fa09376f3576728dbe4eb2e71fbd464568fac" Oct 03 17:30:51 crc kubenswrapper[4744]: I1003 17:30:51.892325 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:30:51 crc kubenswrapper[4744]: E1003 17:30:51.894366 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:31:05 crc kubenswrapper[4744]: I1003 17:31:05.892383 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:31:05 crc kubenswrapper[4744]: E1003 17:31:05.894403 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:31:20 crc kubenswrapper[4744]: I1003 17:31:20.896087 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:31:20 crc kubenswrapper[4744]: E1003 17:31:20.901465 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:31:32 crc kubenswrapper[4744]: I1003 17:31:32.891762 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:31:32 crc kubenswrapper[4744]: E1003 17:31:32.892528 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:31:47 crc kubenswrapper[4744]: I1003 17:31:47.892038 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:31:47 crc kubenswrapper[4744]: E1003 17:31:47.892740 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:31:58 crc kubenswrapper[4744]: I1003 17:31:58.892973 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:31:58 crc kubenswrapper[4744]: E1003 17:31:58.893975 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:32:07 crc kubenswrapper[4744]: I1003 17:32:07.211249 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tm929"] Oct 03 17:32:07 crc kubenswrapper[4744]: E1003 17:32:07.212439 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e978d1e-b25a-4c13-a138-4bdc309ca419" containerName="collect-profiles" Oct 03 17:32:07 crc kubenswrapper[4744]: I1003 17:32:07.212457 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e978d1e-b25a-4c13-a138-4bdc309ca419" containerName="collect-profiles" Oct 03 17:32:07 crc kubenswrapper[4744]: I1003 17:32:07.212716 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e978d1e-b25a-4c13-a138-4bdc309ca419" containerName="collect-profiles" Oct 03 17:32:07 crc kubenswrapper[4744]: I1003 17:32:07.217234 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:07 crc kubenswrapper[4744]: I1003 17:32:07.222810 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tm929"] Oct 03 17:32:07 crc kubenswrapper[4744]: I1003 17:32:07.311303 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09329625-8ae1-463b-b347-ea75e394ab85-utilities\") pod \"redhat-operators-tm929\" (UID: \"09329625-8ae1-463b-b347-ea75e394ab85\") " pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:07 crc kubenswrapper[4744]: I1003 17:32:07.311363 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrq96\" (UniqueName: \"kubernetes.io/projected/09329625-8ae1-463b-b347-ea75e394ab85-kube-api-access-mrq96\") pod \"redhat-operators-tm929\" (UID: \"09329625-8ae1-463b-b347-ea75e394ab85\") " pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:07 crc kubenswrapper[4744]: I1003 17:32:07.311457 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09329625-8ae1-463b-b347-ea75e394ab85-catalog-content\") pod \"redhat-operators-tm929\" (UID: \"09329625-8ae1-463b-b347-ea75e394ab85\") " pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:07 crc kubenswrapper[4744]: I1003 17:32:07.413232 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrq96\" (UniqueName: \"kubernetes.io/projected/09329625-8ae1-463b-b347-ea75e394ab85-kube-api-access-mrq96\") pod \"redhat-operators-tm929\" (UID: \"09329625-8ae1-463b-b347-ea75e394ab85\") " pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:07 crc kubenswrapper[4744]: I1003 17:32:07.413407 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09329625-8ae1-463b-b347-ea75e394ab85-catalog-content\") pod \"redhat-operators-tm929\" (UID: \"09329625-8ae1-463b-b347-ea75e394ab85\") " pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:07 crc kubenswrapper[4744]: I1003 17:32:07.413660 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09329625-8ae1-463b-b347-ea75e394ab85-utilities\") pod \"redhat-operators-tm929\" (UID: \"09329625-8ae1-463b-b347-ea75e394ab85\") " pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:07 crc kubenswrapper[4744]: I1003 17:32:07.414746 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09329625-8ae1-463b-b347-ea75e394ab85-utilities\") pod \"redhat-operators-tm929\" (UID: \"09329625-8ae1-463b-b347-ea75e394ab85\") " pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:07 crc kubenswrapper[4744]: I1003 17:32:07.416304 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09329625-8ae1-463b-b347-ea75e394ab85-catalog-content\") pod \"redhat-operators-tm929\" (UID: \"09329625-8ae1-463b-b347-ea75e394ab85\") " pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:07 crc kubenswrapper[4744]: I1003 17:32:07.436902 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrq96\" (UniqueName: \"kubernetes.io/projected/09329625-8ae1-463b-b347-ea75e394ab85-kube-api-access-mrq96\") pod \"redhat-operators-tm929\" (UID: \"09329625-8ae1-463b-b347-ea75e394ab85\") " pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:07 crc kubenswrapper[4744]: I1003 17:32:07.554461 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:08 crc kubenswrapper[4744]: I1003 17:32:08.060167 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tm929"] Oct 03 17:32:08 crc kubenswrapper[4744]: I1003 17:32:08.322033 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tm929" event={"ID":"09329625-8ae1-463b-b347-ea75e394ab85","Type":"ContainerStarted","Data":"e748dee2de8386639da22edf5aa41a02bb624d229293b4317c68250b0f6f6b30"} Oct 03 17:32:08 crc kubenswrapper[4744]: I1003 17:32:08.322082 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tm929" event={"ID":"09329625-8ae1-463b-b347-ea75e394ab85","Type":"ContainerStarted","Data":"822512f197e98f73ededc8298f2f36b5c237dcd0a38473485401725a9aafd54f"} Oct 03 17:32:08 crc kubenswrapper[4744]: I1003 17:32:08.323848 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 17:32:09 crc kubenswrapper[4744]: I1003 17:32:09.331658 4744 generic.go:334] "Generic (PLEG): container finished" podID="09329625-8ae1-463b-b347-ea75e394ab85" containerID="e748dee2de8386639da22edf5aa41a02bb624d229293b4317c68250b0f6f6b30" exitCode=0 Oct 03 17:32:09 crc kubenswrapper[4744]: I1003 17:32:09.331759 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tm929" event={"ID":"09329625-8ae1-463b-b347-ea75e394ab85","Type":"ContainerDied","Data":"e748dee2de8386639da22edf5aa41a02bb624d229293b4317c68250b0f6f6b30"} Oct 03 17:32:09 crc kubenswrapper[4744]: I1003 17:32:09.331950 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tm929" event={"ID":"09329625-8ae1-463b-b347-ea75e394ab85","Type":"ContainerStarted","Data":"7f1e476e11e5571941ac08a2d0b56b5d30937bce09860040b9f56694b70e5724"} Oct 03 17:32:12 crc kubenswrapper[4744]: I1003 17:32:12.357720 4744 generic.go:334] "Generic (PLEG): container finished" podID="09329625-8ae1-463b-b347-ea75e394ab85" containerID="7f1e476e11e5571941ac08a2d0b56b5d30937bce09860040b9f56694b70e5724" exitCode=0 Oct 03 17:32:12 crc kubenswrapper[4744]: I1003 17:32:12.357804 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tm929" event={"ID":"09329625-8ae1-463b-b347-ea75e394ab85","Type":"ContainerDied","Data":"7f1e476e11e5571941ac08a2d0b56b5d30937bce09860040b9f56694b70e5724"} Oct 03 17:32:12 crc kubenswrapper[4744]: I1003 17:32:12.892716 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:32:13 crc kubenswrapper[4744]: I1003 17:32:13.368102 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"ddc9e1b56446985655c0d63f1a691e89b48b1bef23c7ab851fd75082097ae2d5"} Oct 03 17:32:13 crc kubenswrapper[4744]: I1003 17:32:13.370295 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tm929" event={"ID":"09329625-8ae1-463b-b347-ea75e394ab85","Type":"ContainerStarted","Data":"4bc62c30e1b0e5a68e213aa5561460007d8b55e06000d5017a5a8b292f90c9ea"} Oct 03 17:32:13 crc kubenswrapper[4744]: I1003 17:32:13.410733 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tm929" podStartSLOduration=1.968961031 podStartE2EDuration="6.410717968s" podCreationTimestamp="2025-10-03 17:32:07 +0000 UTC" firstStartedPulling="2025-10-03 17:32:08.323649289 +0000 UTC m=+4054.603525185" lastFinishedPulling="2025-10-03 17:32:12.765406226 +0000 UTC m=+4059.045282122" observedRunningTime="2025-10-03 17:32:13.408291468 +0000 UTC m=+4059.688167364" watchObservedRunningTime="2025-10-03 17:32:13.410717968 +0000 UTC m=+4059.690593864" Oct 03 17:32:17 crc kubenswrapper[4744]: I1003 17:32:17.555160 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:17 crc kubenswrapper[4744]: I1003 17:32:17.555829 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:17 crc kubenswrapper[4744]: I1003 17:32:17.608559 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:18 crc kubenswrapper[4744]: I1003 17:32:18.490487 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:18 crc kubenswrapper[4744]: I1003 17:32:18.540603 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tm929"] Oct 03 17:32:20 crc kubenswrapper[4744]: I1003 17:32:20.441587 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tm929" podUID="09329625-8ae1-463b-b347-ea75e394ab85" containerName="registry-server" containerID="cri-o://4bc62c30e1b0e5a68e213aa5561460007d8b55e06000d5017a5a8b292f90c9ea" gracePeriod=2 Oct 03 17:32:21 crc kubenswrapper[4744]: I1003 17:32:21.455324 4744 generic.go:334] "Generic (PLEG): container finished" podID="09329625-8ae1-463b-b347-ea75e394ab85" containerID="4bc62c30e1b0e5a68e213aa5561460007d8b55e06000d5017a5a8b292f90c9ea" exitCode=0 Oct 03 17:32:21 crc kubenswrapper[4744]: I1003 17:32:21.455415 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tm929" event={"ID":"09329625-8ae1-463b-b347-ea75e394ab85","Type":"ContainerDied","Data":"4bc62c30e1b0e5a68e213aa5561460007d8b55e06000d5017a5a8b292f90c9ea"} Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.256125 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.371479 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09329625-8ae1-463b-b347-ea75e394ab85-catalog-content\") pod \"09329625-8ae1-463b-b347-ea75e394ab85\" (UID: \"09329625-8ae1-463b-b347-ea75e394ab85\") " Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.371611 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09329625-8ae1-463b-b347-ea75e394ab85-utilities\") pod \"09329625-8ae1-463b-b347-ea75e394ab85\" (UID: \"09329625-8ae1-463b-b347-ea75e394ab85\") " Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.371672 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrq96\" (UniqueName: \"kubernetes.io/projected/09329625-8ae1-463b-b347-ea75e394ab85-kube-api-access-mrq96\") pod \"09329625-8ae1-463b-b347-ea75e394ab85\" (UID: \"09329625-8ae1-463b-b347-ea75e394ab85\") " Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.372727 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09329625-8ae1-463b-b347-ea75e394ab85-utilities" (OuterVolumeSpecName: "utilities") pod "09329625-8ae1-463b-b347-ea75e394ab85" (UID: "09329625-8ae1-463b-b347-ea75e394ab85"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.377963 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09329625-8ae1-463b-b347-ea75e394ab85-kube-api-access-mrq96" (OuterVolumeSpecName: "kube-api-access-mrq96") pod "09329625-8ae1-463b-b347-ea75e394ab85" (UID: "09329625-8ae1-463b-b347-ea75e394ab85"). InnerVolumeSpecName "kube-api-access-mrq96". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.468678 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tm929" event={"ID":"09329625-8ae1-463b-b347-ea75e394ab85","Type":"ContainerDied","Data":"822512f197e98f73ededc8298f2f36b5c237dcd0a38473485401725a9aafd54f"} Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.468734 4744 scope.go:117] "RemoveContainer" containerID="4bc62c30e1b0e5a68e213aa5561460007d8b55e06000d5017a5a8b292f90c9ea" Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.468790 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tm929" Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.471954 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09329625-8ae1-463b-b347-ea75e394ab85-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09329625-8ae1-463b-b347-ea75e394ab85" (UID: "09329625-8ae1-463b-b347-ea75e394ab85"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.475001 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09329625-8ae1-463b-b347-ea75e394ab85-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.475044 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09329625-8ae1-463b-b347-ea75e394ab85-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.475057 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrq96\" (UniqueName: \"kubernetes.io/projected/09329625-8ae1-463b-b347-ea75e394ab85-kube-api-access-mrq96\") on node \"crc\" DevicePath \"\"" Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.497754 4744 scope.go:117] "RemoveContainer" containerID="7f1e476e11e5571941ac08a2d0b56b5d30937bce09860040b9f56694b70e5724" Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.533942 4744 scope.go:117] "RemoveContainer" containerID="e748dee2de8386639da22edf5aa41a02bb624d229293b4317c68250b0f6f6b30" Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.811956 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tm929"] Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.820264 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tm929"] Oct 03 17:32:22 crc kubenswrapper[4744]: I1003 17:32:22.904343 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09329625-8ae1-463b-b347-ea75e394ab85" path="/var/lib/kubelet/pods/09329625-8ae1-463b-b347-ea75e394ab85/volumes" Oct 03 17:33:25 crc kubenswrapper[4744]: I1003 17:33:25.963199 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7cldm"] Oct 03 17:33:25 crc kubenswrapper[4744]: E1003 17:33:25.964378 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09329625-8ae1-463b-b347-ea75e394ab85" containerName="registry-server" Oct 03 17:33:25 crc kubenswrapper[4744]: I1003 17:33:25.964395 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="09329625-8ae1-463b-b347-ea75e394ab85" containerName="registry-server" Oct 03 17:33:25 crc kubenswrapper[4744]: E1003 17:33:25.964408 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09329625-8ae1-463b-b347-ea75e394ab85" containerName="extract-content" Oct 03 17:33:25 crc kubenswrapper[4744]: I1003 17:33:25.964415 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="09329625-8ae1-463b-b347-ea75e394ab85" containerName="extract-content" Oct 03 17:33:25 crc kubenswrapper[4744]: E1003 17:33:25.964433 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09329625-8ae1-463b-b347-ea75e394ab85" containerName="extract-utilities" Oct 03 17:33:25 crc kubenswrapper[4744]: I1003 17:33:25.964441 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="09329625-8ae1-463b-b347-ea75e394ab85" containerName="extract-utilities" Oct 03 17:33:25 crc kubenswrapper[4744]: I1003 17:33:25.964666 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="09329625-8ae1-463b-b347-ea75e394ab85" containerName="registry-server" Oct 03 17:33:25 crc kubenswrapper[4744]: I1003 17:33:25.966007 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:25 crc kubenswrapper[4744]: I1003 17:33:25.980986 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cldm"] Oct 03 17:33:26 crc kubenswrapper[4744]: I1003 17:33:26.135160 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92b5e555-b146-4717-8cfc-e45b216f7563-catalog-content\") pod \"redhat-marketplace-7cldm\" (UID: \"92b5e555-b146-4717-8cfc-e45b216f7563\") " pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:26 crc kubenswrapper[4744]: I1003 17:33:26.135246 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rj59\" (UniqueName: \"kubernetes.io/projected/92b5e555-b146-4717-8cfc-e45b216f7563-kube-api-access-6rj59\") pod \"redhat-marketplace-7cldm\" (UID: \"92b5e555-b146-4717-8cfc-e45b216f7563\") " pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:26 crc kubenswrapper[4744]: I1003 17:33:26.135415 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92b5e555-b146-4717-8cfc-e45b216f7563-utilities\") pod \"redhat-marketplace-7cldm\" (UID: \"92b5e555-b146-4717-8cfc-e45b216f7563\") " pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:26 crc kubenswrapper[4744]: I1003 17:33:26.236456 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92b5e555-b146-4717-8cfc-e45b216f7563-utilities\") pod \"redhat-marketplace-7cldm\" (UID: \"92b5e555-b146-4717-8cfc-e45b216f7563\") " pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:26 crc kubenswrapper[4744]: I1003 17:33:26.236550 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92b5e555-b146-4717-8cfc-e45b216f7563-catalog-content\") pod \"redhat-marketplace-7cldm\" (UID: \"92b5e555-b146-4717-8cfc-e45b216f7563\") " pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:26 crc kubenswrapper[4744]: I1003 17:33:26.236594 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rj59\" (UniqueName: \"kubernetes.io/projected/92b5e555-b146-4717-8cfc-e45b216f7563-kube-api-access-6rj59\") pod \"redhat-marketplace-7cldm\" (UID: \"92b5e555-b146-4717-8cfc-e45b216f7563\") " pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:26 crc kubenswrapper[4744]: I1003 17:33:26.237255 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92b5e555-b146-4717-8cfc-e45b216f7563-utilities\") pod \"redhat-marketplace-7cldm\" (UID: \"92b5e555-b146-4717-8cfc-e45b216f7563\") " pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:26 crc kubenswrapper[4744]: I1003 17:33:26.237332 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92b5e555-b146-4717-8cfc-e45b216f7563-catalog-content\") pod \"redhat-marketplace-7cldm\" (UID: \"92b5e555-b146-4717-8cfc-e45b216f7563\") " pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:26 crc kubenswrapper[4744]: I1003 17:33:26.255655 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rj59\" (UniqueName: \"kubernetes.io/projected/92b5e555-b146-4717-8cfc-e45b216f7563-kube-api-access-6rj59\") pod \"redhat-marketplace-7cldm\" (UID: \"92b5e555-b146-4717-8cfc-e45b216f7563\") " pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:26 crc kubenswrapper[4744]: I1003 17:33:26.309834 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:26 crc kubenswrapper[4744]: I1003 17:33:26.839584 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cldm"] Oct 03 17:33:27 crc kubenswrapper[4744]: I1003 17:33:27.048127 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cldm" event={"ID":"92b5e555-b146-4717-8cfc-e45b216f7563","Type":"ContainerStarted","Data":"14a7debb9d321dd18b59d6a0d3c5627d1b98ac0da621eba2c8137bec127cfec2"} Oct 03 17:33:28 crc kubenswrapper[4744]: I1003 17:33:28.060832 4744 generic.go:334] "Generic (PLEG): container finished" podID="92b5e555-b146-4717-8cfc-e45b216f7563" containerID="160ce48041fe6382efae2692085648d18cbbbf2cefd43dc5c9e71d4cb4aaf778" exitCode=0 Oct 03 17:33:28 crc kubenswrapper[4744]: I1003 17:33:28.061181 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cldm" event={"ID":"92b5e555-b146-4717-8cfc-e45b216f7563","Type":"ContainerDied","Data":"160ce48041fe6382efae2692085648d18cbbbf2cefd43dc5c9e71d4cb4aaf778"} Oct 03 17:33:29 crc kubenswrapper[4744]: I1003 17:33:29.072867 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cldm" event={"ID":"92b5e555-b146-4717-8cfc-e45b216f7563","Type":"ContainerStarted","Data":"11f9a59b01cae4141b674e71c55bf88192519071dade6ba1d9b9040cc3eba5f0"} Oct 03 17:33:30 crc kubenswrapper[4744]: I1003 17:33:30.083206 4744 generic.go:334] "Generic (PLEG): container finished" podID="92b5e555-b146-4717-8cfc-e45b216f7563" containerID="11f9a59b01cae4141b674e71c55bf88192519071dade6ba1d9b9040cc3eba5f0" exitCode=0 Oct 03 17:33:30 crc kubenswrapper[4744]: I1003 17:33:30.083241 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cldm" event={"ID":"92b5e555-b146-4717-8cfc-e45b216f7563","Type":"ContainerDied","Data":"11f9a59b01cae4141b674e71c55bf88192519071dade6ba1d9b9040cc3eba5f0"} Oct 03 17:33:31 crc kubenswrapper[4744]: I1003 17:33:31.093373 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cldm" event={"ID":"92b5e555-b146-4717-8cfc-e45b216f7563","Type":"ContainerStarted","Data":"2e297d4d105e0bcf4e2a62b3595a84f755160ff354ab03da5dc1f3854be1067b"} Oct 03 17:33:31 crc kubenswrapper[4744]: I1003 17:33:31.115495 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7cldm" podStartSLOduration=3.546165275 podStartE2EDuration="6.115470114s" podCreationTimestamp="2025-10-03 17:33:25 +0000 UTC" firstStartedPulling="2025-10-03 17:33:28.06829308 +0000 UTC m=+4134.348168976" lastFinishedPulling="2025-10-03 17:33:30.637597909 +0000 UTC m=+4136.917473815" observedRunningTime="2025-10-03 17:33:31.110698096 +0000 UTC m=+4137.390574002" watchObservedRunningTime="2025-10-03 17:33:31.115470114 +0000 UTC m=+4137.395346030" Oct 03 17:33:36 crc kubenswrapper[4744]: I1003 17:33:36.310684 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:36 crc kubenswrapper[4744]: I1003 17:33:36.311765 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:36 crc kubenswrapper[4744]: I1003 17:33:36.554918 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:37 crc kubenswrapper[4744]: I1003 17:33:37.215365 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:37 crc kubenswrapper[4744]: I1003 17:33:37.274262 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cldm"] Oct 03 17:33:39 crc kubenswrapper[4744]: I1003 17:33:39.165227 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7cldm" podUID="92b5e555-b146-4717-8cfc-e45b216f7563" containerName="registry-server" containerID="cri-o://2e297d4d105e0bcf4e2a62b3595a84f755160ff354ab03da5dc1f3854be1067b" gracePeriod=2 Oct 03 17:33:39 crc kubenswrapper[4744]: I1003 17:33:39.737803 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:39 crc kubenswrapper[4744]: I1003 17:33:39.927727 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rj59\" (UniqueName: \"kubernetes.io/projected/92b5e555-b146-4717-8cfc-e45b216f7563-kube-api-access-6rj59\") pod \"92b5e555-b146-4717-8cfc-e45b216f7563\" (UID: \"92b5e555-b146-4717-8cfc-e45b216f7563\") " Oct 03 17:33:39 crc kubenswrapper[4744]: I1003 17:33:39.928075 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92b5e555-b146-4717-8cfc-e45b216f7563-utilities\") pod \"92b5e555-b146-4717-8cfc-e45b216f7563\" (UID: \"92b5e555-b146-4717-8cfc-e45b216f7563\") " Oct 03 17:33:39 crc kubenswrapper[4744]: I1003 17:33:39.928247 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92b5e555-b146-4717-8cfc-e45b216f7563-catalog-content\") pod \"92b5e555-b146-4717-8cfc-e45b216f7563\" (UID: \"92b5e555-b146-4717-8cfc-e45b216f7563\") " Oct 03 17:33:39 crc kubenswrapper[4744]: I1003 17:33:39.929645 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92b5e555-b146-4717-8cfc-e45b216f7563-utilities" (OuterVolumeSpecName: "utilities") pod "92b5e555-b146-4717-8cfc-e45b216f7563" (UID: "92b5e555-b146-4717-8cfc-e45b216f7563"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:33:39 crc kubenswrapper[4744]: I1003 17:33:39.943963 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92b5e555-b146-4717-8cfc-e45b216f7563-kube-api-access-6rj59" (OuterVolumeSpecName: "kube-api-access-6rj59") pod "92b5e555-b146-4717-8cfc-e45b216f7563" (UID: "92b5e555-b146-4717-8cfc-e45b216f7563"). InnerVolumeSpecName "kube-api-access-6rj59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:33:39 crc kubenswrapper[4744]: I1003 17:33:39.950726 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92b5e555-b146-4717-8cfc-e45b216f7563-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92b5e555-b146-4717-8cfc-e45b216f7563" (UID: "92b5e555-b146-4717-8cfc-e45b216f7563"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.031452 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92b5e555-b146-4717-8cfc-e45b216f7563-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.031935 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92b5e555-b146-4717-8cfc-e45b216f7563-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.031953 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rj59\" (UniqueName: \"kubernetes.io/projected/92b5e555-b146-4717-8cfc-e45b216f7563-kube-api-access-6rj59\") on node \"crc\" DevicePath \"\"" Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.177052 4744 generic.go:334] "Generic (PLEG): container finished" podID="92b5e555-b146-4717-8cfc-e45b216f7563" containerID="2e297d4d105e0bcf4e2a62b3595a84f755160ff354ab03da5dc1f3854be1067b" exitCode=0 Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.177090 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cldm" event={"ID":"92b5e555-b146-4717-8cfc-e45b216f7563","Type":"ContainerDied","Data":"2e297d4d105e0bcf4e2a62b3595a84f755160ff354ab03da5dc1f3854be1067b"} Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.177116 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cldm" event={"ID":"92b5e555-b146-4717-8cfc-e45b216f7563","Type":"ContainerDied","Data":"14a7debb9d321dd18b59d6a0d3c5627d1b98ac0da621eba2c8137bec127cfec2"} Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.177132 4744 scope.go:117] "RemoveContainer" containerID="2e297d4d105e0bcf4e2a62b3595a84f755160ff354ab03da5dc1f3854be1067b" Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.177131 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cldm" Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.207670 4744 scope.go:117] "RemoveContainer" containerID="11f9a59b01cae4141b674e71c55bf88192519071dade6ba1d9b9040cc3eba5f0" Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.232934 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cldm"] Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.241256 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cldm"] Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.248273 4744 scope.go:117] "RemoveContainer" containerID="160ce48041fe6382efae2692085648d18cbbbf2cefd43dc5c9e71d4cb4aaf778" Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.294618 4744 scope.go:117] "RemoveContainer" containerID="2e297d4d105e0bcf4e2a62b3595a84f755160ff354ab03da5dc1f3854be1067b" Oct 03 17:33:40 crc kubenswrapper[4744]: E1003 17:33:40.295053 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e297d4d105e0bcf4e2a62b3595a84f755160ff354ab03da5dc1f3854be1067b\": container with ID starting with 2e297d4d105e0bcf4e2a62b3595a84f755160ff354ab03da5dc1f3854be1067b not found: ID does not exist" containerID="2e297d4d105e0bcf4e2a62b3595a84f755160ff354ab03da5dc1f3854be1067b" Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.295089 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e297d4d105e0bcf4e2a62b3595a84f755160ff354ab03da5dc1f3854be1067b"} err="failed to get container status \"2e297d4d105e0bcf4e2a62b3595a84f755160ff354ab03da5dc1f3854be1067b\": rpc error: code = NotFound desc = could not find container \"2e297d4d105e0bcf4e2a62b3595a84f755160ff354ab03da5dc1f3854be1067b\": container with ID starting with 2e297d4d105e0bcf4e2a62b3595a84f755160ff354ab03da5dc1f3854be1067b not found: ID does not exist" Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.295111 4744 scope.go:117] "RemoveContainer" containerID="11f9a59b01cae4141b674e71c55bf88192519071dade6ba1d9b9040cc3eba5f0" Oct 03 17:33:40 crc kubenswrapper[4744]: E1003 17:33:40.295328 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11f9a59b01cae4141b674e71c55bf88192519071dade6ba1d9b9040cc3eba5f0\": container with ID starting with 11f9a59b01cae4141b674e71c55bf88192519071dade6ba1d9b9040cc3eba5f0 not found: ID does not exist" containerID="11f9a59b01cae4141b674e71c55bf88192519071dade6ba1d9b9040cc3eba5f0" Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.295349 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11f9a59b01cae4141b674e71c55bf88192519071dade6ba1d9b9040cc3eba5f0"} err="failed to get container status \"11f9a59b01cae4141b674e71c55bf88192519071dade6ba1d9b9040cc3eba5f0\": rpc error: code = NotFound desc = could not find container \"11f9a59b01cae4141b674e71c55bf88192519071dade6ba1d9b9040cc3eba5f0\": container with ID starting with 11f9a59b01cae4141b674e71c55bf88192519071dade6ba1d9b9040cc3eba5f0 not found: ID does not exist" Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.295363 4744 scope.go:117] "RemoveContainer" containerID="160ce48041fe6382efae2692085648d18cbbbf2cefd43dc5c9e71d4cb4aaf778" Oct 03 17:33:40 crc kubenswrapper[4744]: E1003 17:33:40.295633 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"160ce48041fe6382efae2692085648d18cbbbf2cefd43dc5c9e71d4cb4aaf778\": container with ID starting with 160ce48041fe6382efae2692085648d18cbbbf2cefd43dc5c9e71d4cb4aaf778 not found: ID does not exist" containerID="160ce48041fe6382efae2692085648d18cbbbf2cefd43dc5c9e71d4cb4aaf778" Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.295654 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"160ce48041fe6382efae2692085648d18cbbbf2cefd43dc5c9e71d4cb4aaf778"} err="failed to get container status \"160ce48041fe6382efae2692085648d18cbbbf2cefd43dc5c9e71d4cb4aaf778\": rpc error: code = NotFound desc = could not find container \"160ce48041fe6382efae2692085648d18cbbbf2cefd43dc5c9e71d4cb4aaf778\": container with ID starting with 160ce48041fe6382efae2692085648d18cbbbf2cefd43dc5c9e71d4cb4aaf778 not found: ID does not exist" Oct 03 17:33:40 crc kubenswrapper[4744]: I1003 17:33:40.909835 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92b5e555-b146-4717-8cfc-e45b216f7563" path="/var/lib/kubelet/pods/92b5e555-b146-4717-8cfc-e45b216f7563/volumes" Oct 03 17:34:34 crc kubenswrapper[4744]: I1003 17:34:34.668879 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:34:34 crc kubenswrapper[4744]: I1003 17:34:34.669479 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:35:04 crc kubenswrapper[4744]: I1003 17:35:04.668469 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:35:04 crc kubenswrapper[4744]: I1003 17:35:04.669175 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:35:34 crc kubenswrapper[4744]: I1003 17:35:34.668359 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:35:34 crc kubenswrapper[4744]: I1003 17:35:34.669043 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:35:34 crc kubenswrapper[4744]: I1003 17:35:34.669113 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 17:35:34 crc kubenswrapper[4744]: I1003 17:35:34.671859 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ddc9e1b56446985655c0d63f1a691e89b48b1bef23c7ab851fd75082097ae2d5"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 17:35:34 crc kubenswrapper[4744]: I1003 17:35:34.672144 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://ddc9e1b56446985655c0d63f1a691e89b48b1bef23c7ab851fd75082097ae2d5" gracePeriod=600 Oct 03 17:35:35 crc kubenswrapper[4744]: I1003 17:35:35.271408 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="ddc9e1b56446985655c0d63f1a691e89b48b1bef23c7ab851fd75082097ae2d5" exitCode=0 Oct 03 17:35:35 crc kubenswrapper[4744]: I1003 17:35:35.271456 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"ddc9e1b56446985655c0d63f1a691e89b48b1bef23c7ab851fd75082097ae2d5"} Oct 03 17:35:35 crc kubenswrapper[4744]: I1003 17:35:35.271990 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537"} Oct 03 17:35:35 crc kubenswrapper[4744]: I1003 17:35:35.272015 4744 scope.go:117] "RemoveContainer" containerID="6af8abeea44692eb6af83e81a87886dc7a2e0775b55b04bb11df5e435b4c10a6" Oct 03 17:36:48 crc kubenswrapper[4744]: I1003 17:36:48.822849 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7c5r9"] Oct 03 17:36:48 crc kubenswrapper[4744]: E1003 17:36:48.823775 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92b5e555-b146-4717-8cfc-e45b216f7563" containerName="extract-utilities" Oct 03 17:36:48 crc kubenswrapper[4744]: I1003 17:36:48.823790 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="92b5e555-b146-4717-8cfc-e45b216f7563" containerName="extract-utilities" Oct 03 17:36:48 crc kubenswrapper[4744]: E1003 17:36:48.823815 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92b5e555-b146-4717-8cfc-e45b216f7563" containerName="extract-content" Oct 03 17:36:48 crc kubenswrapper[4744]: I1003 17:36:48.823823 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="92b5e555-b146-4717-8cfc-e45b216f7563" containerName="extract-content" Oct 03 17:36:48 crc kubenswrapper[4744]: E1003 17:36:48.823841 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92b5e555-b146-4717-8cfc-e45b216f7563" containerName="registry-server" Oct 03 17:36:48 crc kubenswrapper[4744]: I1003 17:36:48.823850 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="92b5e555-b146-4717-8cfc-e45b216f7563" containerName="registry-server" Oct 03 17:36:48 crc kubenswrapper[4744]: I1003 17:36:48.824117 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="92b5e555-b146-4717-8cfc-e45b216f7563" containerName="registry-server" Oct 03 17:36:48 crc kubenswrapper[4744]: I1003 17:36:48.826328 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:36:48 crc kubenswrapper[4744]: I1003 17:36:48.841918 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7c5r9"] Oct 03 17:36:48 crc kubenswrapper[4744]: I1003 17:36:48.899903 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gp2kj\" (UniqueName: \"kubernetes.io/projected/11b5f801-40c3-44da-b4c0-2ccd34392592-kube-api-access-gp2kj\") pod \"certified-operators-7c5r9\" (UID: \"11b5f801-40c3-44da-b4c0-2ccd34392592\") " pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:36:48 crc kubenswrapper[4744]: I1003 17:36:48.900025 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11b5f801-40c3-44da-b4c0-2ccd34392592-utilities\") pod \"certified-operators-7c5r9\" (UID: \"11b5f801-40c3-44da-b4c0-2ccd34392592\") " pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:36:48 crc kubenswrapper[4744]: I1003 17:36:48.900098 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11b5f801-40c3-44da-b4c0-2ccd34392592-catalog-content\") pod \"certified-operators-7c5r9\" (UID: \"11b5f801-40c3-44da-b4c0-2ccd34392592\") " pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:36:49 crc kubenswrapper[4744]: I1003 17:36:49.001382 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gp2kj\" (UniqueName: \"kubernetes.io/projected/11b5f801-40c3-44da-b4c0-2ccd34392592-kube-api-access-gp2kj\") pod \"certified-operators-7c5r9\" (UID: \"11b5f801-40c3-44da-b4c0-2ccd34392592\") " pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:36:49 crc kubenswrapper[4744]: I1003 17:36:49.001560 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11b5f801-40c3-44da-b4c0-2ccd34392592-utilities\") pod \"certified-operators-7c5r9\" (UID: \"11b5f801-40c3-44da-b4c0-2ccd34392592\") " pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:36:49 crc kubenswrapper[4744]: I1003 17:36:49.001623 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11b5f801-40c3-44da-b4c0-2ccd34392592-catalog-content\") pod \"certified-operators-7c5r9\" (UID: \"11b5f801-40c3-44da-b4c0-2ccd34392592\") " pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:36:49 crc kubenswrapper[4744]: I1003 17:36:49.002136 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11b5f801-40c3-44da-b4c0-2ccd34392592-catalog-content\") pod \"certified-operators-7c5r9\" (UID: \"11b5f801-40c3-44da-b4c0-2ccd34392592\") " pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:36:49 crc kubenswrapper[4744]: I1003 17:36:49.002178 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11b5f801-40c3-44da-b4c0-2ccd34392592-utilities\") pod \"certified-operators-7c5r9\" (UID: \"11b5f801-40c3-44da-b4c0-2ccd34392592\") " pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:36:49 crc kubenswrapper[4744]: I1003 17:36:49.030630 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gp2kj\" (UniqueName: \"kubernetes.io/projected/11b5f801-40c3-44da-b4c0-2ccd34392592-kube-api-access-gp2kj\") pod \"certified-operators-7c5r9\" (UID: \"11b5f801-40c3-44da-b4c0-2ccd34392592\") " pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:36:49 crc kubenswrapper[4744]: I1003 17:36:49.155568 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:36:49 crc kubenswrapper[4744]: I1003 17:36:49.735102 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7c5r9"] Oct 03 17:36:49 crc kubenswrapper[4744]: W1003 17:36:49.741901 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11b5f801_40c3_44da_b4c0_2ccd34392592.slice/crio-e0a28cf9187e3248bddf3bc958f7bdd5ec906c7474d24869893b308c6b64c661 WatchSource:0}: Error finding container e0a28cf9187e3248bddf3bc958f7bdd5ec906c7474d24869893b308c6b64c661: Status 404 returned error can't find the container with id e0a28cf9187e3248bddf3bc958f7bdd5ec906c7474d24869893b308c6b64c661 Oct 03 17:36:50 crc kubenswrapper[4744]: I1003 17:36:50.021393 4744 generic.go:334] "Generic (PLEG): container finished" podID="11b5f801-40c3-44da-b4c0-2ccd34392592" containerID="ff275d1fdc49c3a24ad6b6f90ba69e14257cfb133af92e53c9a018e3d0a1b9db" exitCode=0 Oct 03 17:36:50 crc kubenswrapper[4744]: I1003 17:36:50.021440 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7c5r9" event={"ID":"11b5f801-40c3-44da-b4c0-2ccd34392592","Type":"ContainerDied","Data":"ff275d1fdc49c3a24ad6b6f90ba69e14257cfb133af92e53c9a018e3d0a1b9db"} Oct 03 17:36:50 crc kubenswrapper[4744]: I1003 17:36:50.021481 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7c5r9" event={"ID":"11b5f801-40c3-44da-b4c0-2ccd34392592","Type":"ContainerStarted","Data":"e0a28cf9187e3248bddf3bc958f7bdd5ec906c7474d24869893b308c6b64c661"} Oct 03 17:36:52 crc kubenswrapper[4744]: I1003 17:36:52.041319 4744 generic.go:334] "Generic (PLEG): container finished" podID="11b5f801-40c3-44da-b4c0-2ccd34392592" containerID="45c42581496671054d49f6147d1e8bff65018630565e721dfd5ff7115c8548be" exitCode=0 Oct 03 17:36:52 crc kubenswrapper[4744]: I1003 17:36:52.041398 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7c5r9" event={"ID":"11b5f801-40c3-44da-b4c0-2ccd34392592","Type":"ContainerDied","Data":"45c42581496671054d49f6147d1e8bff65018630565e721dfd5ff7115c8548be"} Oct 03 17:36:53 crc kubenswrapper[4744]: I1003 17:36:53.052086 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7c5r9" event={"ID":"11b5f801-40c3-44da-b4c0-2ccd34392592","Type":"ContainerStarted","Data":"41a3586cae3bb90a38564b885cf32a9df5d3645930e3cbec81996e956a930a97"} Oct 03 17:36:53 crc kubenswrapper[4744]: I1003 17:36:53.081797 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7c5r9" podStartSLOduration=2.650210651 podStartE2EDuration="5.081772507s" podCreationTimestamp="2025-10-03 17:36:48 +0000 UTC" firstStartedPulling="2025-10-03 17:36:50.025120369 +0000 UTC m=+4336.304996265" lastFinishedPulling="2025-10-03 17:36:52.456682225 +0000 UTC m=+4338.736558121" observedRunningTime="2025-10-03 17:36:53.072169871 +0000 UTC m=+4339.352045777" watchObservedRunningTime="2025-10-03 17:36:53.081772507 +0000 UTC m=+4339.361648403" Oct 03 17:36:59 crc kubenswrapper[4744]: I1003 17:36:59.157198 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:36:59 crc kubenswrapper[4744]: I1003 17:36:59.157893 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:36:59 crc kubenswrapper[4744]: I1003 17:36:59.220709 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:37:00 crc kubenswrapper[4744]: I1003 17:37:00.164357 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:37:00 crc kubenswrapper[4744]: I1003 17:37:00.210556 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7c5r9"] Oct 03 17:37:02 crc kubenswrapper[4744]: I1003 17:37:02.136597 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7c5r9" podUID="11b5f801-40c3-44da-b4c0-2ccd34392592" containerName="registry-server" containerID="cri-o://41a3586cae3bb90a38564b885cf32a9df5d3645930e3cbec81996e956a930a97" gracePeriod=2 Oct 03 17:37:02 crc kubenswrapper[4744]: I1003 17:37:02.799704 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:37:02 crc kubenswrapper[4744]: I1003 17:37:02.888574 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11b5f801-40c3-44da-b4c0-2ccd34392592-catalog-content\") pod \"11b5f801-40c3-44da-b4c0-2ccd34392592\" (UID: \"11b5f801-40c3-44da-b4c0-2ccd34392592\") " Oct 03 17:37:02 crc kubenswrapper[4744]: I1003 17:37:02.888698 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11b5f801-40c3-44da-b4c0-2ccd34392592-utilities\") pod \"11b5f801-40c3-44da-b4c0-2ccd34392592\" (UID: \"11b5f801-40c3-44da-b4c0-2ccd34392592\") " Oct 03 17:37:02 crc kubenswrapper[4744]: I1003 17:37:02.888853 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gp2kj\" (UniqueName: \"kubernetes.io/projected/11b5f801-40c3-44da-b4c0-2ccd34392592-kube-api-access-gp2kj\") pod \"11b5f801-40c3-44da-b4c0-2ccd34392592\" (UID: \"11b5f801-40c3-44da-b4c0-2ccd34392592\") " Oct 03 17:37:02 crc kubenswrapper[4744]: I1003 17:37:02.889785 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11b5f801-40c3-44da-b4c0-2ccd34392592-utilities" (OuterVolumeSpecName: "utilities") pod "11b5f801-40c3-44da-b4c0-2ccd34392592" (UID: "11b5f801-40c3-44da-b4c0-2ccd34392592"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:37:02 crc kubenswrapper[4744]: I1003 17:37:02.896049 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11b5f801-40c3-44da-b4c0-2ccd34392592-kube-api-access-gp2kj" (OuterVolumeSpecName: "kube-api-access-gp2kj") pod "11b5f801-40c3-44da-b4c0-2ccd34392592" (UID: "11b5f801-40c3-44da-b4c0-2ccd34392592"). InnerVolumeSpecName "kube-api-access-gp2kj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:37:02 crc kubenswrapper[4744]: I1003 17:37:02.944921 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11b5f801-40c3-44da-b4c0-2ccd34392592-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "11b5f801-40c3-44da-b4c0-2ccd34392592" (UID: "11b5f801-40c3-44da-b4c0-2ccd34392592"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:37:02 crc kubenswrapper[4744]: I1003 17:37:02.994167 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11b5f801-40c3-44da-b4c0-2ccd34392592-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:37:02 crc kubenswrapper[4744]: I1003 17:37:02.994199 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11b5f801-40c3-44da-b4c0-2ccd34392592-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:37:02 crc kubenswrapper[4744]: I1003 17:37:02.994215 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gp2kj\" (UniqueName: \"kubernetes.io/projected/11b5f801-40c3-44da-b4c0-2ccd34392592-kube-api-access-gp2kj\") on node \"crc\" DevicePath \"\"" Oct 03 17:37:03 crc kubenswrapper[4744]: I1003 17:37:03.155742 4744 generic.go:334] "Generic (PLEG): container finished" podID="11b5f801-40c3-44da-b4c0-2ccd34392592" containerID="41a3586cae3bb90a38564b885cf32a9df5d3645930e3cbec81996e956a930a97" exitCode=0 Oct 03 17:37:03 crc kubenswrapper[4744]: I1003 17:37:03.155784 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7c5r9" event={"ID":"11b5f801-40c3-44da-b4c0-2ccd34392592","Type":"ContainerDied","Data":"41a3586cae3bb90a38564b885cf32a9df5d3645930e3cbec81996e956a930a97"} Oct 03 17:37:03 crc kubenswrapper[4744]: I1003 17:37:03.155819 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7c5r9" event={"ID":"11b5f801-40c3-44da-b4c0-2ccd34392592","Type":"ContainerDied","Data":"e0a28cf9187e3248bddf3bc958f7bdd5ec906c7474d24869893b308c6b64c661"} Oct 03 17:37:03 crc kubenswrapper[4744]: I1003 17:37:03.155841 4744 scope.go:117] "RemoveContainer" containerID="41a3586cae3bb90a38564b885cf32a9df5d3645930e3cbec81996e956a930a97" Oct 03 17:37:03 crc kubenswrapper[4744]: I1003 17:37:03.155951 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7c5r9" Oct 03 17:37:03 crc kubenswrapper[4744]: I1003 17:37:03.190162 4744 scope.go:117] "RemoveContainer" containerID="45c42581496671054d49f6147d1e8bff65018630565e721dfd5ff7115c8548be" Oct 03 17:37:03 crc kubenswrapper[4744]: I1003 17:37:03.194489 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7c5r9"] Oct 03 17:37:03 crc kubenswrapper[4744]: I1003 17:37:03.204524 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7c5r9"] Oct 03 17:37:03 crc kubenswrapper[4744]: I1003 17:37:03.839414 4744 scope.go:117] "RemoveContainer" containerID="ff275d1fdc49c3a24ad6b6f90ba69e14257cfb133af92e53c9a018e3d0a1b9db" Oct 03 17:37:03 crc kubenswrapper[4744]: I1003 17:37:03.880381 4744 scope.go:117] "RemoveContainer" containerID="41a3586cae3bb90a38564b885cf32a9df5d3645930e3cbec81996e956a930a97" Oct 03 17:37:03 crc kubenswrapper[4744]: E1003 17:37:03.881186 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41a3586cae3bb90a38564b885cf32a9df5d3645930e3cbec81996e956a930a97\": container with ID starting with 41a3586cae3bb90a38564b885cf32a9df5d3645930e3cbec81996e956a930a97 not found: ID does not exist" containerID="41a3586cae3bb90a38564b885cf32a9df5d3645930e3cbec81996e956a930a97" Oct 03 17:37:03 crc kubenswrapper[4744]: I1003 17:37:03.881238 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41a3586cae3bb90a38564b885cf32a9df5d3645930e3cbec81996e956a930a97"} err="failed to get container status \"41a3586cae3bb90a38564b885cf32a9df5d3645930e3cbec81996e956a930a97\": rpc error: code = NotFound desc = could not find container \"41a3586cae3bb90a38564b885cf32a9df5d3645930e3cbec81996e956a930a97\": container with ID starting with 41a3586cae3bb90a38564b885cf32a9df5d3645930e3cbec81996e956a930a97 not found: ID does not exist" Oct 03 17:37:03 crc kubenswrapper[4744]: I1003 17:37:03.881265 4744 scope.go:117] "RemoveContainer" containerID="45c42581496671054d49f6147d1e8bff65018630565e721dfd5ff7115c8548be" Oct 03 17:37:03 crc kubenswrapper[4744]: E1003 17:37:03.881747 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45c42581496671054d49f6147d1e8bff65018630565e721dfd5ff7115c8548be\": container with ID starting with 45c42581496671054d49f6147d1e8bff65018630565e721dfd5ff7115c8548be not found: ID does not exist" containerID="45c42581496671054d49f6147d1e8bff65018630565e721dfd5ff7115c8548be" Oct 03 17:37:03 crc kubenswrapper[4744]: I1003 17:37:03.881772 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45c42581496671054d49f6147d1e8bff65018630565e721dfd5ff7115c8548be"} err="failed to get container status \"45c42581496671054d49f6147d1e8bff65018630565e721dfd5ff7115c8548be\": rpc error: code = NotFound desc = could not find container \"45c42581496671054d49f6147d1e8bff65018630565e721dfd5ff7115c8548be\": container with ID starting with 45c42581496671054d49f6147d1e8bff65018630565e721dfd5ff7115c8548be not found: ID does not exist" Oct 03 17:37:03 crc kubenswrapper[4744]: I1003 17:37:03.881789 4744 scope.go:117] "RemoveContainer" containerID="ff275d1fdc49c3a24ad6b6f90ba69e14257cfb133af92e53c9a018e3d0a1b9db" Oct 03 17:37:03 crc kubenswrapper[4744]: E1003 17:37:03.882088 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff275d1fdc49c3a24ad6b6f90ba69e14257cfb133af92e53c9a018e3d0a1b9db\": container with ID starting with ff275d1fdc49c3a24ad6b6f90ba69e14257cfb133af92e53c9a018e3d0a1b9db not found: ID does not exist" containerID="ff275d1fdc49c3a24ad6b6f90ba69e14257cfb133af92e53c9a018e3d0a1b9db" Oct 03 17:37:03 crc kubenswrapper[4744]: I1003 17:37:03.882110 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff275d1fdc49c3a24ad6b6f90ba69e14257cfb133af92e53c9a018e3d0a1b9db"} err="failed to get container status \"ff275d1fdc49c3a24ad6b6f90ba69e14257cfb133af92e53c9a018e3d0a1b9db\": rpc error: code = NotFound desc = could not find container \"ff275d1fdc49c3a24ad6b6f90ba69e14257cfb133af92e53c9a018e3d0a1b9db\": container with ID starting with ff275d1fdc49c3a24ad6b6f90ba69e14257cfb133af92e53c9a018e3d0a1b9db not found: ID does not exist" Oct 03 17:37:04 crc kubenswrapper[4744]: I1003 17:37:04.904864 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11b5f801-40c3-44da-b4c0-2ccd34392592" path="/var/lib/kubelet/pods/11b5f801-40c3-44da-b4c0-2ccd34392592/volumes" Oct 03 17:38:04 crc kubenswrapper[4744]: I1003 17:38:04.668393 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:38:04 crc kubenswrapper[4744]: I1003 17:38:04.668950 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:38:34 crc kubenswrapper[4744]: I1003 17:38:34.668990 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:38:34 crc kubenswrapper[4744]: I1003 17:38:34.669776 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:39:04 crc kubenswrapper[4744]: I1003 17:39:04.668025 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:39:04 crc kubenswrapper[4744]: I1003 17:39:04.668763 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:39:04 crc kubenswrapper[4744]: I1003 17:39:04.668828 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 17:39:04 crc kubenswrapper[4744]: I1003 17:39:04.670032 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 17:39:04 crc kubenswrapper[4744]: I1003 17:39:04.670165 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" gracePeriod=600 Oct 03 17:39:04 crc kubenswrapper[4744]: E1003 17:39:04.798669 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:39:05 crc kubenswrapper[4744]: I1003 17:39:05.362449 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" exitCode=0 Oct 03 17:39:05 crc kubenswrapper[4744]: I1003 17:39:05.362545 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537"} Oct 03 17:39:05 crc kubenswrapper[4744]: I1003 17:39:05.362829 4744 scope.go:117] "RemoveContainer" containerID="ddc9e1b56446985655c0d63f1a691e89b48b1bef23c7ab851fd75082097ae2d5" Oct 03 17:39:05 crc kubenswrapper[4744]: I1003 17:39:05.363712 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:39:05 crc kubenswrapper[4744]: E1003 17:39:05.364133 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:39:18 crc kubenswrapper[4744]: I1003 17:39:18.892183 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:39:18 crc kubenswrapper[4744]: E1003 17:39:18.893134 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:39:29 crc kubenswrapper[4744]: I1003 17:39:29.891960 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:39:29 crc kubenswrapper[4744]: E1003 17:39:29.892698 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:39:44 crc kubenswrapper[4744]: I1003 17:39:44.908809 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:39:44 crc kubenswrapper[4744]: E1003 17:39:44.910345 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:39:55 crc kubenswrapper[4744]: I1003 17:39:55.892269 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:39:55 crc kubenswrapper[4744]: E1003 17:39:55.893054 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:40:08 crc kubenswrapper[4744]: I1003 17:40:08.894267 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:40:08 crc kubenswrapper[4744]: E1003 17:40:08.895265 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:40:23 crc kubenswrapper[4744]: I1003 17:40:23.892279 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:40:23 crc kubenswrapper[4744]: E1003 17:40:23.893833 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:40:37 crc kubenswrapper[4744]: I1003 17:40:37.892460 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:40:37 crc kubenswrapper[4744]: E1003 17:40:37.893278 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:40:48 crc kubenswrapper[4744]: I1003 17:40:48.893373 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:40:48 crc kubenswrapper[4744]: E1003 17:40:48.894477 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:40:59 crc kubenswrapper[4744]: I1003 17:40:59.891615 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:40:59 crc kubenswrapper[4744]: E1003 17:40:59.892784 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:41:06 crc kubenswrapper[4744]: I1003 17:41:06.606326 4744 generic.go:334] "Generic (PLEG): container finished" podID="595f783e-db00-49b2-968f-79163001ffdd" containerID="81d520babfe8bb321dedaa86a51f6868965891969a4eba83624f528fedc8b468" exitCode=0 Oct 03 17:41:06 crc kubenswrapper[4744]: I1003 17:41:06.606436 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"595f783e-db00-49b2-968f-79163001ffdd","Type":"ContainerDied","Data":"81d520babfe8bb321dedaa86a51f6868965891969a4eba83624f528fedc8b468"} Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.034735 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.122224 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/595f783e-db00-49b2-968f-79163001ffdd-test-operator-ephemeral-workdir\") pod \"595f783e-db00-49b2-968f-79163001ffdd\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.122292 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-openstack-config-secret\") pod \"595f783e-db00-49b2-968f-79163001ffdd\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.122369 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"595f783e-db00-49b2-968f-79163001ffdd\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.122423 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/595f783e-db00-49b2-968f-79163001ffdd-test-operator-ephemeral-temporary\") pod \"595f783e-db00-49b2-968f-79163001ffdd\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.122479 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/595f783e-db00-49b2-968f-79163001ffdd-config-data\") pod \"595f783e-db00-49b2-968f-79163001ffdd\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.122559 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/595f783e-db00-49b2-968f-79163001ffdd-openstack-config\") pod \"595f783e-db00-49b2-968f-79163001ffdd\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.122593 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-ca-certs\") pod \"595f783e-db00-49b2-968f-79163001ffdd\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.122623 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-ssh-key\") pod \"595f783e-db00-49b2-968f-79163001ffdd\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.122688 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b797f\" (UniqueName: \"kubernetes.io/projected/595f783e-db00-49b2-968f-79163001ffdd-kube-api-access-b797f\") pod \"595f783e-db00-49b2-968f-79163001ffdd\" (UID: \"595f783e-db00-49b2-968f-79163001ffdd\") " Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.125720 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/595f783e-db00-49b2-968f-79163001ffdd-config-data" (OuterVolumeSpecName: "config-data") pod "595f783e-db00-49b2-968f-79163001ffdd" (UID: "595f783e-db00-49b2-968f-79163001ffdd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.125891 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/595f783e-db00-49b2-968f-79163001ffdd-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "595f783e-db00-49b2-968f-79163001ffdd" (UID: "595f783e-db00-49b2-968f-79163001ffdd"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.133525 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "595f783e-db00-49b2-968f-79163001ffdd" (UID: "595f783e-db00-49b2-968f-79163001ffdd"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.133679 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/595f783e-db00-49b2-968f-79163001ffdd-kube-api-access-b797f" (OuterVolumeSpecName: "kube-api-access-b797f") pod "595f783e-db00-49b2-968f-79163001ffdd" (UID: "595f783e-db00-49b2-968f-79163001ffdd"). InnerVolumeSpecName "kube-api-access-b797f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.136628 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/595f783e-db00-49b2-968f-79163001ffdd-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "595f783e-db00-49b2-968f-79163001ffdd" (UID: "595f783e-db00-49b2-968f-79163001ffdd"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.161232 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "595f783e-db00-49b2-968f-79163001ffdd" (UID: "595f783e-db00-49b2-968f-79163001ffdd"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.166484 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "595f783e-db00-49b2-968f-79163001ffdd" (UID: "595f783e-db00-49b2-968f-79163001ffdd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.168016 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "595f783e-db00-49b2-968f-79163001ffdd" (UID: "595f783e-db00-49b2-968f-79163001ffdd"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.199940 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/595f783e-db00-49b2-968f-79163001ffdd-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "595f783e-db00-49b2-968f-79163001ffdd" (UID: "595f783e-db00-49b2-968f-79163001ffdd"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.224626 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.224702 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.224717 4744 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/595f783e-db00-49b2-968f-79163001ffdd-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.224728 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/595f783e-db00-49b2-968f-79163001ffdd-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.224737 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/595f783e-db00-49b2-968f-79163001ffdd-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.224749 4744 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.224757 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/595f783e-db00-49b2-968f-79163001ffdd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.224766 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b797f\" (UniqueName: \"kubernetes.io/projected/595f783e-db00-49b2-968f-79163001ffdd-kube-api-access-b797f\") on node \"crc\" DevicePath \"\"" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.224774 4744 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/595f783e-db00-49b2-968f-79163001ffdd-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.246667 4744 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.326526 4744 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.625735 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"595f783e-db00-49b2-968f-79163001ffdd","Type":"ContainerDied","Data":"ce59043602b7ff62ea8d1aa8c885a946755ab4381b792d8f8b9ba6e442ed3b2f"} Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.625792 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce59043602b7ff62ea8d1aa8c885a946755ab4381b792d8f8b9ba6e442ed3b2f" Oct 03 17:41:08 crc kubenswrapper[4744]: I1003 17:41:08.625813 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.552169 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 03 17:41:10 crc kubenswrapper[4744]: E1003 17:41:10.553293 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b5f801-40c3-44da-b4c0-2ccd34392592" containerName="extract-utilities" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.553310 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b5f801-40c3-44da-b4c0-2ccd34392592" containerName="extract-utilities" Oct 03 17:41:10 crc kubenswrapper[4744]: E1003 17:41:10.553334 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b5f801-40c3-44da-b4c0-2ccd34392592" containerName="registry-server" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.553342 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b5f801-40c3-44da-b4c0-2ccd34392592" containerName="registry-server" Oct 03 17:41:10 crc kubenswrapper[4744]: E1003 17:41:10.553360 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="595f783e-db00-49b2-968f-79163001ffdd" containerName="tempest-tests-tempest-tests-runner" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.553370 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="595f783e-db00-49b2-968f-79163001ffdd" containerName="tempest-tests-tempest-tests-runner" Oct 03 17:41:10 crc kubenswrapper[4744]: E1003 17:41:10.553411 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b5f801-40c3-44da-b4c0-2ccd34392592" containerName="extract-content" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.553419 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b5f801-40c3-44da-b4c0-2ccd34392592" containerName="extract-content" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.553675 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="11b5f801-40c3-44da-b4c0-2ccd34392592" containerName="registry-server" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.553696 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="595f783e-db00-49b2-968f-79163001ffdd" containerName="tempest-tests-tempest-tests-runner" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.554579 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.568543 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.679837 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"812d60e3-44e3-4f22-9afa-448f0ac216ef\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.679896 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjlpq\" (UniqueName: \"kubernetes.io/projected/812d60e3-44e3-4f22-9afa-448f0ac216ef-kube-api-access-mjlpq\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"812d60e3-44e3-4f22-9afa-448f0ac216ef\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.781851 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"812d60e3-44e3-4f22-9afa-448f0ac216ef\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.781914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjlpq\" (UniqueName: \"kubernetes.io/projected/812d60e3-44e3-4f22-9afa-448f0ac216ef-kube-api-access-mjlpq\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"812d60e3-44e3-4f22-9afa-448f0ac216ef\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.782830 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"812d60e3-44e3-4f22-9afa-448f0ac216ef\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.812178 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjlpq\" (UniqueName: \"kubernetes.io/projected/812d60e3-44e3-4f22-9afa-448f0ac216ef-kube-api-access-mjlpq\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"812d60e3-44e3-4f22-9afa-448f0ac216ef\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.814164 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"812d60e3-44e3-4f22-9afa-448f0ac216ef\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 03 17:41:10 crc kubenswrapper[4744]: I1003 17:41:10.885791 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 03 17:41:11 crc kubenswrapper[4744]: I1003 17:41:11.360595 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 03 17:41:11 crc kubenswrapper[4744]: I1003 17:41:11.373067 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 17:41:11 crc kubenswrapper[4744]: I1003 17:41:11.664274 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"812d60e3-44e3-4f22-9afa-448f0ac216ef","Type":"ContainerStarted","Data":"d051bd5c93c27a777bbc2a4e10f0dffcbfafd110f145b182b06810d79821f2fa"} Oct 03 17:41:12 crc kubenswrapper[4744]: I1003 17:41:12.683664 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"812d60e3-44e3-4f22-9afa-448f0ac216ef","Type":"ContainerStarted","Data":"b08813aefeb047e94015d34e421327d123a7e9c4f0c9397246539c185f2758f9"} Oct 03 17:41:12 crc kubenswrapper[4744]: I1003 17:41:12.715119 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.720995547 podStartE2EDuration="2.715088262s" podCreationTimestamp="2025-10-03 17:41:10 +0000 UTC" firstStartedPulling="2025-10-03 17:41:11.372774517 +0000 UTC m=+4597.652650413" lastFinishedPulling="2025-10-03 17:41:12.366867192 +0000 UTC m=+4598.646743128" observedRunningTime="2025-10-03 17:41:12.703746012 +0000 UTC m=+4598.983621918" watchObservedRunningTime="2025-10-03 17:41:12.715088262 +0000 UTC m=+4598.994964168" Oct 03 17:41:12 crc kubenswrapper[4744]: I1003 17:41:12.893686 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:41:12 crc kubenswrapper[4744]: E1003 17:41:12.894138 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:41:27 crc kubenswrapper[4744]: I1003 17:41:27.892885 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:41:27 crc kubenswrapper[4744]: E1003 17:41:27.894177 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:41:30 crc kubenswrapper[4744]: I1003 17:41:30.408697 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mhn28/must-gather-pzkn4"] Oct 03 17:41:30 crc kubenswrapper[4744]: I1003 17:41:30.411167 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/must-gather-pzkn4" Oct 03 17:41:30 crc kubenswrapper[4744]: I1003 17:41:30.423515 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mhn28"/"kube-root-ca.crt" Oct 03 17:41:30 crc kubenswrapper[4744]: I1003 17:41:30.423545 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mhn28/must-gather-pzkn4"] Oct 03 17:41:30 crc kubenswrapper[4744]: I1003 17:41:30.423648 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mhn28"/"default-dockercfg-b677s" Oct 03 17:41:30 crc kubenswrapper[4744]: I1003 17:41:30.427134 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mhn28"/"openshift-service-ca.crt" Oct 03 17:41:30 crc kubenswrapper[4744]: I1003 17:41:30.532799 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmmdk\" (UniqueName: \"kubernetes.io/projected/c7c5232f-a750-4320-95b6-df7b9866babe-kube-api-access-dmmdk\") pod \"must-gather-pzkn4\" (UID: \"c7c5232f-a750-4320-95b6-df7b9866babe\") " pod="openshift-must-gather-mhn28/must-gather-pzkn4" Oct 03 17:41:30 crc kubenswrapper[4744]: I1003 17:41:30.532984 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c7c5232f-a750-4320-95b6-df7b9866babe-must-gather-output\") pod \"must-gather-pzkn4\" (UID: \"c7c5232f-a750-4320-95b6-df7b9866babe\") " pod="openshift-must-gather-mhn28/must-gather-pzkn4" Oct 03 17:41:30 crc kubenswrapper[4744]: I1003 17:41:30.634591 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmmdk\" (UniqueName: \"kubernetes.io/projected/c7c5232f-a750-4320-95b6-df7b9866babe-kube-api-access-dmmdk\") pod \"must-gather-pzkn4\" (UID: \"c7c5232f-a750-4320-95b6-df7b9866babe\") " pod="openshift-must-gather-mhn28/must-gather-pzkn4" Oct 03 17:41:30 crc kubenswrapper[4744]: I1003 17:41:30.634703 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c7c5232f-a750-4320-95b6-df7b9866babe-must-gather-output\") pod \"must-gather-pzkn4\" (UID: \"c7c5232f-a750-4320-95b6-df7b9866babe\") " pod="openshift-must-gather-mhn28/must-gather-pzkn4" Oct 03 17:41:30 crc kubenswrapper[4744]: I1003 17:41:30.635145 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c7c5232f-a750-4320-95b6-df7b9866babe-must-gather-output\") pod \"must-gather-pzkn4\" (UID: \"c7c5232f-a750-4320-95b6-df7b9866babe\") " pod="openshift-must-gather-mhn28/must-gather-pzkn4" Oct 03 17:41:30 crc kubenswrapper[4744]: I1003 17:41:30.664517 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmmdk\" (UniqueName: \"kubernetes.io/projected/c7c5232f-a750-4320-95b6-df7b9866babe-kube-api-access-dmmdk\") pod \"must-gather-pzkn4\" (UID: \"c7c5232f-a750-4320-95b6-df7b9866babe\") " pod="openshift-must-gather-mhn28/must-gather-pzkn4" Oct 03 17:41:30 crc kubenswrapper[4744]: I1003 17:41:30.743618 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/must-gather-pzkn4" Oct 03 17:41:31 crc kubenswrapper[4744]: I1003 17:41:31.198484 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mhn28/must-gather-pzkn4"] Oct 03 17:41:31 crc kubenswrapper[4744]: I1003 17:41:31.235030 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2fqg7"] Oct 03 17:41:31 crc kubenswrapper[4744]: I1003 17:41:31.237741 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:31 crc kubenswrapper[4744]: I1003 17:41:31.248091 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2fqg7"] Oct 03 17:41:31 crc kubenswrapper[4744]: I1003 17:41:31.347664 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b05e14d1-6e82-4126-9c82-b3f21def4077-catalog-content\") pod \"community-operators-2fqg7\" (UID: \"b05e14d1-6e82-4126-9c82-b3f21def4077\") " pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:31 crc kubenswrapper[4744]: I1003 17:41:31.347719 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b05e14d1-6e82-4126-9c82-b3f21def4077-utilities\") pod \"community-operators-2fqg7\" (UID: \"b05e14d1-6e82-4126-9c82-b3f21def4077\") " pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:31 crc kubenswrapper[4744]: I1003 17:41:31.347886 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cj68\" (UniqueName: \"kubernetes.io/projected/b05e14d1-6e82-4126-9c82-b3f21def4077-kube-api-access-4cj68\") pod \"community-operators-2fqg7\" (UID: \"b05e14d1-6e82-4126-9c82-b3f21def4077\") " pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:31 crc kubenswrapper[4744]: I1003 17:41:31.450066 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cj68\" (UniqueName: \"kubernetes.io/projected/b05e14d1-6e82-4126-9c82-b3f21def4077-kube-api-access-4cj68\") pod \"community-operators-2fqg7\" (UID: \"b05e14d1-6e82-4126-9c82-b3f21def4077\") " pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:31 crc kubenswrapper[4744]: I1003 17:41:31.450164 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b05e14d1-6e82-4126-9c82-b3f21def4077-catalog-content\") pod \"community-operators-2fqg7\" (UID: \"b05e14d1-6e82-4126-9c82-b3f21def4077\") " pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:31 crc kubenswrapper[4744]: I1003 17:41:31.450206 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b05e14d1-6e82-4126-9c82-b3f21def4077-utilities\") pod \"community-operators-2fqg7\" (UID: \"b05e14d1-6e82-4126-9c82-b3f21def4077\") " pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:31 crc kubenswrapper[4744]: I1003 17:41:31.450783 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b05e14d1-6e82-4126-9c82-b3f21def4077-utilities\") pod \"community-operators-2fqg7\" (UID: \"b05e14d1-6e82-4126-9c82-b3f21def4077\") " pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:31 crc kubenswrapper[4744]: I1003 17:41:31.451346 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b05e14d1-6e82-4126-9c82-b3f21def4077-catalog-content\") pod \"community-operators-2fqg7\" (UID: \"b05e14d1-6e82-4126-9c82-b3f21def4077\") " pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:31 crc kubenswrapper[4744]: I1003 17:41:31.470263 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cj68\" (UniqueName: \"kubernetes.io/projected/b05e14d1-6e82-4126-9c82-b3f21def4077-kube-api-access-4cj68\") pod \"community-operators-2fqg7\" (UID: \"b05e14d1-6e82-4126-9c82-b3f21def4077\") " pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:31 crc kubenswrapper[4744]: I1003 17:41:31.586526 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:31 crc kubenswrapper[4744]: I1003 17:41:31.900529 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mhn28/must-gather-pzkn4" event={"ID":"c7c5232f-a750-4320-95b6-df7b9866babe","Type":"ContainerStarted","Data":"2cdd8ef5d3e1b951a54baf8b2365e884bb61c6126125d1bbb6dbe8b29ca8c7fc"} Oct 03 17:41:32 crc kubenswrapper[4744]: I1003 17:41:32.152354 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2fqg7"] Oct 03 17:41:32 crc kubenswrapper[4744]: W1003 17:41:32.522691 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb05e14d1_6e82_4126_9c82_b3f21def4077.slice/crio-3ee9215835a4699944ec1d2be5823299def7f8ea8618c47c5d6a5abb262eed73 WatchSource:0}: Error finding container 3ee9215835a4699944ec1d2be5823299def7f8ea8618c47c5d6a5abb262eed73: Status 404 returned error can't find the container with id 3ee9215835a4699944ec1d2be5823299def7f8ea8618c47c5d6a5abb262eed73 Oct 03 17:41:32 crc kubenswrapper[4744]: I1003 17:41:32.927531 4744 generic.go:334] "Generic (PLEG): container finished" podID="b05e14d1-6e82-4126-9c82-b3f21def4077" containerID="87b379473a2b6ca8bdd7ab04cc98f3787b1d4004b62040de739c7ff0db0fcf76" exitCode=0 Oct 03 17:41:32 crc kubenswrapper[4744]: I1003 17:41:32.927577 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2fqg7" event={"ID":"b05e14d1-6e82-4126-9c82-b3f21def4077","Type":"ContainerDied","Data":"87b379473a2b6ca8bdd7ab04cc98f3787b1d4004b62040de739c7ff0db0fcf76"} Oct 03 17:41:32 crc kubenswrapper[4744]: I1003 17:41:32.927612 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2fqg7" event={"ID":"b05e14d1-6e82-4126-9c82-b3f21def4077","Type":"ContainerStarted","Data":"3ee9215835a4699944ec1d2be5823299def7f8ea8618c47c5d6a5abb262eed73"} Oct 03 17:41:35 crc kubenswrapper[4744]: I1003 17:41:35.963736 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mhn28/must-gather-pzkn4" event={"ID":"c7c5232f-a750-4320-95b6-df7b9866babe","Type":"ContainerStarted","Data":"a2a8f55a9d23bbf8785a3ba76c418a62f7df52b1f839687bc6660f09b9a8edce"} Oct 03 17:41:35 crc kubenswrapper[4744]: I1003 17:41:35.964128 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mhn28/must-gather-pzkn4" event={"ID":"c7c5232f-a750-4320-95b6-df7b9866babe","Type":"ContainerStarted","Data":"3bd725223847135cc3e7488666794ed27f33c99e5549007dee5277afe20ca140"} Oct 03 17:41:35 crc kubenswrapper[4744]: I1003 17:41:35.965896 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2fqg7" event={"ID":"b05e14d1-6e82-4126-9c82-b3f21def4077","Type":"ContainerStarted","Data":"787f59f2420b231e8144144ef539b1e9606d0310317350be80a361a5de58d679"} Oct 03 17:41:35 crc kubenswrapper[4744]: I1003 17:41:35.981897 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mhn28/must-gather-pzkn4" podStartSLOduration=1.84209514 podStartE2EDuration="5.981881997s" podCreationTimestamp="2025-10-03 17:41:30 +0000 UTC" firstStartedPulling="2025-10-03 17:41:31.204963171 +0000 UTC m=+4617.484839067" lastFinishedPulling="2025-10-03 17:41:35.344750028 +0000 UTC m=+4621.624625924" observedRunningTime="2025-10-03 17:41:35.980172275 +0000 UTC m=+4622.260048171" watchObservedRunningTime="2025-10-03 17:41:35.981881997 +0000 UTC m=+4622.261757893" Oct 03 17:41:36 crc kubenswrapper[4744]: I1003 17:41:36.975608 4744 generic.go:334] "Generic (PLEG): container finished" podID="b05e14d1-6e82-4126-9c82-b3f21def4077" containerID="787f59f2420b231e8144144ef539b1e9606d0310317350be80a361a5de58d679" exitCode=0 Oct 03 17:41:36 crc kubenswrapper[4744]: I1003 17:41:36.975753 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2fqg7" event={"ID":"b05e14d1-6e82-4126-9c82-b3f21def4077","Type":"ContainerDied","Data":"787f59f2420b231e8144144ef539b1e9606d0310317350be80a361a5de58d679"} Oct 03 17:41:37 crc kubenswrapper[4744]: I1003 17:41:37.987288 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2fqg7" event={"ID":"b05e14d1-6e82-4126-9c82-b3f21def4077","Type":"ContainerStarted","Data":"51ee5020a69231f9b10728f78203d1534c73c540fcec68ea8b48b9f71c9084d8"} Oct 03 17:41:38 crc kubenswrapper[4744]: I1003 17:41:38.010230 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2fqg7" podStartSLOduration=2.406262692 podStartE2EDuration="7.010210246s" podCreationTimestamp="2025-10-03 17:41:31 +0000 UTC" firstStartedPulling="2025-10-03 17:41:32.938875345 +0000 UTC m=+4619.218751241" lastFinishedPulling="2025-10-03 17:41:37.542822899 +0000 UTC m=+4623.822698795" observedRunningTime="2025-10-03 17:41:38.006041093 +0000 UTC m=+4624.285916989" watchObservedRunningTime="2025-10-03 17:41:38.010210246 +0000 UTC m=+4624.290086142" Oct 03 17:41:40 crc kubenswrapper[4744]: I1003 17:41:40.892077 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:41:40 crc kubenswrapper[4744]: E1003 17:41:40.892915 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:41:41 crc kubenswrapper[4744]: I1003 17:41:41.588249 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:41 crc kubenswrapper[4744]: I1003 17:41:41.588775 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:41 crc kubenswrapper[4744]: I1003 17:41:41.612722 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mhn28/crc-debug-6hv6g"] Oct 03 17:41:41 crc kubenswrapper[4744]: I1003 17:41:41.615076 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/crc-debug-6hv6g" Oct 03 17:41:41 crc kubenswrapper[4744]: I1003 17:41:41.658658 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/292f592c-6a56-48eb-acda-07416514c6ab-host\") pod \"crc-debug-6hv6g\" (UID: \"292f592c-6a56-48eb-acda-07416514c6ab\") " pod="openshift-must-gather-mhn28/crc-debug-6hv6g" Oct 03 17:41:41 crc kubenswrapper[4744]: I1003 17:41:41.658976 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl4k2\" (UniqueName: \"kubernetes.io/projected/292f592c-6a56-48eb-acda-07416514c6ab-kube-api-access-kl4k2\") pod \"crc-debug-6hv6g\" (UID: \"292f592c-6a56-48eb-acda-07416514c6ab\") " pod="openshift-must-gather-mhn28/crc-debug-6hv6g" Oct 03 17:41:41 crc kubenswrapper[4744]: I1003 17:41:41.662322 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:41 crc kubenswrapper[4744]: I1003 17:41:41.760549 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl4k2\" (UniqueName: \"kubernetes.io/projected/292f592c-6a56-48eb-acda-07416514c6ab-kube-api-access-kl4k2\") pod \"crc-debug-6hv6g\" (UID: \"292f592c-6a56-48eb-acda-07416514c6ab\") " pod="openshift-must-gather-mhn28/crc-debug-6hv6g" Oct 03 17:41:41 crc kubenswrapper[4744]: I1003 17:41:41.760863 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/292f592c-6a56-48eb-acda-07416514c6ab-host\") pod \"crc-debug-6hv6g\" (UID: \"292f592c-6a56-48eb-acda-07416514c6ab\") " pod="openshift-must-gather-mhn28/crc-debug-6hv6g" Oct 03 17:41:41 crc kubenswrapper[4744]: I1003 17:41:41.760909 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/292f592c-6a56-48eb-acda-07416514c6ab-host\") pod \"crc-debug-6hv6g\" (UID: \"292f592c-6a56-48eb-acda-07416514c6ab\") " pod="openshift-must-gather-mhn28/crc-debug-6hv6g" Oct 03 17:41:41 crc kubenswrapper[4744]: I1003 17:41:41.782541 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl4k2\" (UniqueName: \"kubernetes.io/projected/292f592c-6a56-48eb-acda-07416514c6ab-kube-api-access-kl4k2\") pod \"crc-debug-6hv6g\" (UID: \"292f592c-6a56-48eb-acda-07416514c6ab\") " pod="openshift-must-gather-mhn28/crc-debug-6hv6g" Oct 03 17:41:41 crc kubenswrapper[4744]: I1003 17:41:41.934636 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/crc-debug-6hv6g" Oct 03 17:41:42 crc kubenswrapper[4744]: I1003 17:41:42.021083 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mhn28/crc-debug-6hv6g" event={"ID":"292f592c-6a56-48eb-acda-07416514c6ab","Type":"ContainerStarted","Data":"3d3ca51ed035d3399d6bf21e24eff737975f239a062076bbc360be5712a7c6ca"} Oct 03 17:41:42 crc kubenswrapper[4744]: I1003 17:41:42.073050 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:42 crc kubenswrapper[4744]: I1003 17:41:42.117724 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2fqg7"] Oct 03 17:41:44 crc kubenswrapper[4744]: I1003 17:41:44.037536 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2fqg7" podUID="b05e14d1-6e82-4126-9c82-b3f21def4077" containerName="registry-server" containerID="cri-o://51ee5020a69231f9b10728f78203d1534c73c540fcec68ea8b48b9f71c9084d8" gracePeriod=2 Oct 03 17:41:44 crc kubenswrapper[4744]: I1003 17:41:44.608853 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:44 crc kubenswrapper[4744]: I1003 17:41:44.631203 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b05e14d1-6e82-4126-9c82-b3f21def4077-catalog-content\") pod \"b05e14d1-6e82-4126-9c82-b3f21def4077\" (UID: \"b05e14d1-6e82-4126-9c82-b3f21def4077\") " Oct 03 17:41:44 crc kubenswrapper[4744]: I1003 17:41:44.644947 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cj68\" (UniqueName: \"kubernetes.io/projected/b05e14d1-6e82-4126-9c82-b3f21def4077-kube-api-access-4cj68\") pod \"b05e14d1-6e82-4126-9c82-b3f21def4077\" (UID: \"b05e14d1-6e82-4126-9c82-b3f21def4077\") " Oct 03 17:41:44 crc kubenswrapper[4744]: I1003 17:41:44.645270 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b05e14d1-6e82-4126-9c82-b3f21def4077-utilities\") pod \"b05e14d1-6e82-4126-9c82-b3f21def4077\" (UID: \"b05e14d1-6e82-4126-9c82-b3f21def4077\") " Oct 03 17:41:44 crc kubenswrapper[4744]: I1003 17:41:44.646277 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b05e14d1-6e82-4126-9c82-b3f21def4077-utilities" (OuterVolumeSpecName: "utilities") pod "b05e14d1-6e82-4126-9c82-b3f21def4077" (UID: "b05e14d1-6e82-4126-9c82-b3f21def4077"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:41:44 crc kubenswrapper[4744]: I1003 17:41:44.658087 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b05e14d1-6e82-4126-9c82-b3f21def4077-kube-api-access-4cj68" (OuterVolumeSpecName: "kube-api-access-4cj68") pod "b05e14d1-6e82-4126-9c82-b3f21def4077" (UID: "b05e14d1-6e82-4126-9c82-b3f21def4077"). InnerVolumeSpecName "kube-api-access-4cj68". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:41:44 crc kubenswrapper[4744]: I1003 17:41:44.688395 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b05e14d1-6e82-4126-9c82-b3f21def4077-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b05e14d1-6e82-4126-9c82-b3f21def4077" (UID: "b05e14d1-6e82-4126-9c82-b3f21def4077"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:41:44 crc kubenswrapper[4744]: I1003 17:41:44.747318 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b05e14d1-6e82-4126-9c82-b3f21def4077-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:41:44 crc kubenswrapper[4744]: I1003 17:41:44.747347 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cj68\" (UniqueName: \"kubernetes.io/projected/b05e14d1-6e82-4126-9c82-b3f21def4077-kube-api-access-4cj68\") on node \"crc\" DevicePath \"\"" Oct 03 17:41:44 crc kubenswrapper[4744]: I1003 17:41:44.747359 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b05e14d1-6e82-4126-9c82-b3f21def4077-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:41:45 crc kubenswrapper[4744]: I1003 17:41:45.050096 4744 generic.go:334] "Generic (PLEG): container finished" podID="b05e14d1-6e82-4126-9c82-b3f21def4077" containerID="51ee5020a69231f9b10728f78203d1534c73c540fcec68ea8b48b9f71c9084d8" exitCode=0 Oct 03 17:41:45 crc kubenswrapper[4744]: I1003 17:41:45.050139 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2fqg7" event={"ID":"b05e14d1-6e82-4126-9c82-b3f21def4077","Type":"ContainerDied","Data":"51ee5020a69231f9b10728f78203d1534c73c540fcec68ea8b48b9f71c9084d8"} Oct 03 17:41:45 crc kubenswrapper[4744]: I1003 17:41:45.050165 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2fqg7" event={"ID":"b05e14d1-6e82-4126-9c82-b3f21def4077","Type":"ContainerDied","Data":"3ee9215835a4699944ec1d2be5823299def7f8ea8618c47c5d6a5abb262eed73"} Oct 03 17:41:45 crc kubenswrapper[4744]: I1003 17:41:45.050181 4744 scope.go:117] "RemoveContainer" containerID="51ee5020a69231f9b10728f78203d1534c73c540fcec68ea8b48b9f71c9084d8" Oct 03 17:41:45 crc kubenswrapper[4744]: I1003 17:41:45.050190 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2fqg7" Oct 03 17:41:45 crc kubenswrapper[4744]: I1003 17:41:45.072987 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2fqg7"] Oct 03 17:41:45 crc kubenswrapper[4744]: I1003 17:41:45.082646 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2fqg7"] Oct 03 17:41:45 crc kubenswrapper[4744]: I1003 17:41:45.092572 4744 scope.go:117] "RemoveContainer" containerID="787f59f2420b231e8144144ef539b1e9606d0310317350be80a361a5de58d679" Oct 03 17:41:45 crc kubenswrapper[4744]: I1003 17:41:45.123393 4744 scope.go:117] "RemoveContainer" containerID="87b379473a2b6ca8bdd7ab04cc98f3787b1d4004b62040de739c7ff0db0fcf76" Oct 03 17:41:45 crc kubenswrapper[4744]: I1003 17:41:45.149424 4744 scope.go:117] "RemoveContainer" containerID="51ee5020a69231f9b10728f78203d1534c73c540fcec68ea8b48b9f71c9084d8" Oct 03 17:41:45 crc kubenswrapper[4744]: E1003 17:41:45.149935 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51ee5020a69231f9b10728f78203d1534c73c540fcec68ea8b48b9f71c9084d8\": container with ID starting with 51ee5020a69231f9b10728f78203d1534c73c540fcec68ea8b48b9f71c9084d8 not found: ID does not exist" containerID="51ee5020a69231f9b10728f78203d1534c73c540fcec68ea8b48b9f71c9084d8" Oct 03 17:41:45 crc kubenswrapper[4744]: I1003 17:41:45.149991 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51ee5020a69231f9b10728f78203d1534c73c540fcec68ea8b48b9f71c9084d8"} err="failed to get container status \"51ee5020a69231f9b10728f78203d1534c73c540fcec68ea8b48b9f71c9084d8\": rpc error: code = NotFound desc = could not find container \"51ee5020a69231f9b10728f78203d1534c73c540fcec68ea8b48b9f71c9084d8\": container with ID starting with 51ee5020a69231f9b10728f78203d1534c73c540fcec68ea8b48b9f71c9084d8 not found: ID does not exist" Oct 03 17:41:45 crc kubenswrapper[4744]: I1003 17:41:45.150018 4744 scope.go:117] "RemoveContainer" containerID="787f59f2420b231e8144144ef539b1e9606d0310317350be80a361a5de58d679" Oct 03 17:41:45 crc kubenswrapper[4744]: E1003 17:41:45.150384 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"787f59f2420b231e8144144ef539b1e9606d0310317350be80a361a5de58d679\": container with ID starting with 787f59f2420b231e8144144ef539b1e9606d0310317350be80a361a5de58d679 not found: ID does not exist" containerID="787f59f2420b231e8144144ef539b1e9606d0310317350be80a361a5de58d679" Oct 03 17:41:45 crc kubenswrapper[4744]: I1003 17:41:45.150415 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"787f59f2420b231e8144144ef539b1e9606d0310317350be80a361a5de58d679"} err="failed to get container status \"787f59f2420b231e8144144ef539b1e9606d0310317350be80a361a5de58d679\": rpc error: code = NotFound desc = could not find container \"787f59f2420b231e8144144ef539b1e9606d0310317350be80a361a5de58d679\": container with ID starting with 787f59f2420b231e8144144ef539b1e9606d0310317350be80a361a5de58d679 not found: ID does not exist" Oct 03 17:41:45 crc kubenswrapper[4744]: I1003 17:41:45.150435 4744 scope.go:117] "RemoveContainer" containerID="87b379473a2b6ca8bdd7ab04cc98f3787b1d4004b62040de739c7ff0db0fcf76" Oct 03 17:41:45 crc kubenswrapper[4744]: E1003 17:41:45.151590 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87b379473a2b6ca8bdd7ab04cc98f3787b1d4004b62040de739c7ff0db0fcf76\": container with ID starting with 87b379473a2b6ca8bdd7ab04cc98f3787b1d4004b62040de739c7ff0db0fcf76 not found: ID does not exist" containerID="87b379473a2b6ca8bdd7ab04cc98f3787b1d4004b62040de739c7ff0db0fcf76" Oct 03 17:41:45 crc kubenswrapper[4744]: I1003 17:41:45.151621 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87b379473a2b6ca8bdd7ab04cc98f3787b1d4004b62040de739c7ff0db0fcf76"} err="failed to get container status \"87b379473a2b6ca8bdd7ab04cc98f3787b1d4004b62040de739c7ff0db0fcf76\": rpc error: code = NotFound desc = could not find container \"87b379473a2b6ca8bdd7ab04cc98f3787b1d4004b62040de739c7ff0db0fcf76\": container with ID starting with 87b379473a2b6ca8bdd7ab04cc98f3787b1d4004b62040de739c7ff0db0fcf76 not found: ID does not exist" Oct 03 17:41:46 crc kubenswrapper[4744]: I1003 17:41:46.905549 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b05e14d1-6e82-4126-9c82-b3f21def4077" path="/var/lib/kubelet/pods/b05e14d1-6e82-4126-9c82-b3f21def4077/volumes" Oct 03 17:41:53 crc kubenswrapper[4744]: I1003 17:41:53.123182 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mhn28/crc-debug-6hv6g" event={"ID":"292f592c-6a56-48eb-acda-07416514c6ab","Type":"ContainerStarted","Data":"01b8eec269bfadcd5704da9f37361eaa3d2894284c41385246ea3e7f2e6dd0a6"} Oct 03 17:41:53 crc kubenswrapper[4744]: I1003 17:41:53.161910 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mhn28/crc-debug-6hv6g" podStartSLOduration=1.704359713 podStartE2EDuration="12.161890241s" podCreationTimestamp="2025-10-03 17:41:41 +0000 UTC" firstStartedPulling="2025-10-03 17:41:41.980200969 +0000 UTC m=+4628.260076865" lastFinishedPulling="2025-10-03 17:41:52.437731497 +0000 UTC m=+4638.717607393" observedRunningTime="2025-10-03 17:41:53.160922007 +0000 UTC m=+4639.440797903" watchObservedRunningTime="2025-10-03 17:41:53.161890241 +0000 UTC m=+4639.441766147" Oct 03 17:41:53 crc kubenswrapper[4744]: I1003 17:41:53.892746 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:41:53 crc kubenswrapper[4744]: E1003 17:41:53.893621 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:42:04 crc kubenswrapper[4744]: I1003 17:42:04.904979 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:42:04 crc kubenswrapper[4744]: E1003 17:42:04.906035 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:42:15 crc kubenswrapper[4744]: I1003 17:42:15.892151 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:42:15 crc kubenswrapper[4744]: E1003 17:42:15.894134 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.492167 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wwv64"] Oct 03 17:42:28 crc kubenswrapper[4744]: E1003 17:42:28.493293 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b05e14d1-6e82-4126-9c82-b3f21def4077" containerName="extract-utilities" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.493312 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b05e14d1-6e82-4126-9c82-b3f21def4077" containerName="extract-utilities" Oct 03 17:42:28 crc kubenswrapper[4744]: E1003 17:42:28.493325 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b05e14d1-6e82-4126-9c82-b3f21def4077" containerName="registry-server" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.493333 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b05e14d1-6e82-4126-9c82-b3f21def4077" containerName="registry-server" Oct 03 17:42:28 crc kubenswrapper[4744]: E1003 17:42:28.493357 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b05e14d1-6e82-4126-9c82-b3f21def4077" containerName="extract-content" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.493364 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b05e14d1-6e82-4126-9c82-b3f21def4077" containerName="extract-content" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.493633 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b05e14d1-6e82-4126-9c82-b3f21def4077" containerName="registry-server" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.495314 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.504891 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wwv64"] Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.625940 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-utilities\") pod \"redhat-operators-wwv64\" (UID: \"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8\") " pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.626265 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7tt7\" (UniqueName: \"kubernetes.io/projected/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-kube-api-access-l7tt7\") pod \"redhat-operators-wwv64\" (UID: \"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8\") " pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.626408 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-catalog-content\") pod \"redhat-operators-wwv64\" (UID: \"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8\") " pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.728073 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7tt7\" (UniqueName: \"kubernetes.io/projected/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-kube-api-access-l7tt7\") pod \"redhat-operators-wwv64\" (UID: \"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8\") " pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.728289 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-catalog-content\") pod \"redhat-operators-wwv64\" (UID: \"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8\") " pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.728449 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-utilities\") pod \"redhat-operators-wwv64\" (UID: \"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8\") " pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.728928 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-catalog-content\") pod \"redhat-operators-wwv64\" (UID: \"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8\") " pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.729026 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-utilities\") pod \"redhat-operators-wwv64\" (UID: \"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8\") " pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.753528 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7tt7\" (UniqueName: \"kubernetes.io/projected/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-kube-api-access-l7tt7\") pod \"redhat-operators-wwv64\" (UID: \"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8\") " pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:28 crc kubenswrapper[4744]: I1003 17:42:28.824549 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:29 crc kubenswrapper[4744]: I1003 17:42:29.358202 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wwv64"] Oct 03 17:42:29 crc kubenswrapper[4744]: I1003 17:42:29.471590 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwv64" event={"ID":"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8","Type":"ContainerStarted","Data":"5348f0220836674f4845aae4ce448260c782445257f727ace095ccc9d4eb50e2"} Oct 03 17:42:29 crc kubenswrapper[4744]: I1003 17:42:29.891938 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:42:29 crc kubenswrapper[4744]: E1003 17:42:29.892201 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:42:30 crc kubenswrapper[4744]: I1003 17:42:30.482058 4744 generic.go:334] "Generic (PLEG): container finished" podID="a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8" containerID="99e4c3e77d4edddbe59c22e1f6be9b9efb3e296b5319db43753ec7e80afce58e" exitCode=0 Oct 03 17:42:30 crc kubenswrapper[4744]: I1003 17:42:30.482104 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwv64" event={"ID":"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8","Type":"ContainerDied","Data":"99e4c3e77d4edddbe59c22e1f6be9b9efb3e296b5319db43753ec7e80afce58e"} Oct 03 17:42:31 crc kubenswrapper[4744]: I1003 17:42:31.492768 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwv64" event={"ID":"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8","Type":"ContainerStarted","Data":"29fc529935e8198bc4a2a500009abec66dd6ff756974baee661fff107efcd54d"} Oct 03 17:42:33 crc kubenswrapper[4744]: I1003 17:42:33.514960 4744 generic.go:334] "Generic (PLEG): container finished" podID="a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8" containerID="29fc529935e8198bc4a2a500009abec66dd6ff756974baee661fff107efcd54d" exitCode=0 Oct 03 17:42:33 crc kubenswrapper[4744]: I1003 17:42:33.515075 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwv64" event={"ID":"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8","Type":"ContainerDied","Data":"29fc529935e8198bc4a2a500009abec66dd6ff756974baee661fff107efcd54d"} Oct 03 17:42:34 crc kubenswrapper[4744]: I1003 17:42:34.534214 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwv64" event={"ID":"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8","Type":"ContainerStarted","Data":"4993a613b99c63696bdbefc29e4b02c2886f1750e52f80b79de4dffa60d5975f"} Oct 03 17:42:34 crc kubenswrapper[4744]: I1003 17:42:34.558307 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wwv64" podStartSLOduration=2.923653515 podStartE2EDuration="6.558286368s" podCreationTimestamp="2025-10-03 17:42:28 +0000 UTC" firstStartedPulling="2025-10-03 17:42:30.485479793 +0000 UTC m=+4676.765355679" lastFinishedPulling="2025-10-03 17:42:34.120112636 +0000 UTC m=+4680.399988532" observedRunningTime="2025-10-03 17:42:34.550682288 +0000 UTC m=+4680.830558184" watchObservedRunningTime="2025-10-03 17:42:34.558286368 +0000 UTC m=+4680.838162264" Oct 03 17:42:38 crc kubenswrapper[4744]: I1003 17:42:38.826207 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:38 crc kubenswrapper[4744]: I1003 17:42:38.826884 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:38 crc kubenswrapper[4744]: I1003 17:42:38.878192 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:39 crc kubenswrapper[4744]: I1003 17:42:39.645828 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:39 crc kubenswrapper[4744]: I1003 17:42:39.718978 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wwv64"] Oct 03 17:42:41 crc kubenswrapper[4744]: I1003 17:42:41.602128 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wwv64" podUID="a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8" containerName="registry-server" containerID="cri-o://4993a613b99c63696bdbefc29e4b02c2886f1750e52f80b79de4dffa60d5975f" gracePeriod=2 Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.126702 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.196986 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-catalog-content\") pod \"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8\" (UID: \"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8\") " Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.197069 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-utilities\") pod \"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8\" (UID: \"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8\") " Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.197207 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7tt7\" (UniqueName: \"kubernetes.io/projected/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-kube-api-access-l7tt7\") pod \"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8\" (UID: \"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8\") " Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.197872 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-utilities" (OuterVolumeSpecName: "utilities") pod "a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8" (UID: "a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.203564 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-kube-api-access-l7tt7" (OuterVolumeSpecName: "kube-api-access-l7tt7") pod "a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8" (UID: "a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8"). InnerVolumeSpecName "kube-api-access-l7tt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.296262 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8" (UID: "a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.299787 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.299829 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.299840 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7tt7\" (UniqueName: \"kubernetes.io/projected/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8-kube-api-access-l7tt7\") on node \"crc\" DevicePath \"\"" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.613592 4744 generic.go:334] "Generic (PLEG): container finished" podID="a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8" containerID="4993a613b99c63696bdbefc29e4b02c2886f1750e52f80b79de4dffa60d5975f" exitCode=0 Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.613633 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwv64" event={"ID":"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8","Type":"ContainerDied","Data":"4993a613b99c63696bdbefc29e4b02c2886f1750e52f80b79de4dffa60d5975f"} Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.613665 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wwv64" event={"ID":"a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8","Type":"ContainerDied","Data":"5348f0220836674f4845aae4ce448260c782445257f727ace095ccc9d4eb50e2"} Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.613684 4744 scope.go:117] "RemoveContainer" containerID="4993a613b99c63696bdbefc29e4b02c2886f1750e52f80b79de4dffa60d5975f" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.613680 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wwv64" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.655149 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wwv64"] Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.662661 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wwv64"] Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.663405 4744 scope.go:117] "RemoveContainer" containerID="29fc529935e8198bc4a2a500009abec66dd6ff756974baee661fff107efcd54d" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.686886 4744 scope.go:117] "RemoveContainer" containerID="99e4c3e77d4edddbe59c22e1f6be9b9efb3e296b5319db43753ec7e80afce58e" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.734779 4744 scope.go:117] "RemoveContainer" containerID="4993a613b99c63696bdbefc29e4b02c2886f1750e52f80b79de4dffa60d5975f" Oct 03 17:42:42 crc kubenswrapper[4744]: E1003 17:42:42.735471 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4993a613b99c63696bdbefc29e4b02c2886f1750e52f80b79de4dffa60d5975f\": container with ID starting with 4993a613b99c63696bdbefc29e4b02c2886f1750e52f80b79de4dffa60d5975f not found: ID does not exist" containerID="4993a613b99c63696bdbefc29e4b02c2886f1750e52f80b79de4dffa60d5975f" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.735722 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4993a613b99c63696bdbefc29e4b02c2886f1750e52f80b79de4dffa60d5975f"} err="failed to get container status \"4993a613b99c63696bdbefc29e4b02c2886f1750e52f80b79de4dffa60d5975f\": rpc error: code = NotFound desc = could not find container \"4993a613b99c63696bdbefc29e4b02c2886f1750e52f80b79de4dffa60d5975f\": container with ID starting with 4993a613b99c63696bdbefc29e4b02c2886f1750e52f80b79de4dffa60d5975f not found: ID does not exist" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.735756 4744 scope.go:117] "RemoveContainer" containerID="29fc529935e8198bc4a2a500009abec66dd6ff756974baee661fff107efcd54d" Oct 03 17:42:42 crc kubenswrapper[4744]: E1003 17:42:42.737513 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29fc529935e8198bc4a2a500009abec66dd6ff756974baee661fff107efcd54d\": container with ID starting with 29fc529935e8198bc4a2a500009abec66dd6ff756974baee661fff107efcd54d not found: ID does not exist" containerID="29fc529935e8198bc4a2a500009abec66dd6ff756974baee661fff107efcd54d" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.737593 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29fc529935e8198bc4a2a500009abec66dd6ff756974baee661fff107efcd54d"} err="failed to get container status \"29fc529935e8198bc4a2a500009abec66dd6ff756974baee661fff107efcd54d\": rpc error: code = NotFound desc = could not find container \"29fc529935e8198bc4a2a500009abec66dd6ff756974baee661fff107efcd54d\": container with ID starting with 29fc529935e8198bc4a2a500009abec66dd6ff756974baee661fff107efcd54d not found: ID does not exist" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.737631 4744 scope.go:117] "RemoveContainer" containerID="99e4c3e77d4edddbe59c22e1f6be9b9efb3e296b5319db43753ec7e80afce58e" Oct 03 17:42:42 crc kubenswrapper[4744]: E1003 17:42:42.746770 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99e4c3e77d4edddbe59c22e1f6be9b9efb3e296b5319db43753ec7e80afce58e\": container with ID starting with 99e4c3e77d4edddbe59c22e1f6be9b9efb3e296b5319db43753ec7e80afce58e not found: ID does not exist" containerID="99e4c3e77d4edddbe59c22e1f6be9b9efb3e296b5319db43753ec7e80afce58e" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.746816 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99e4c3e77d4edddbe59c22e1f6be9b9efb3e296b5319db43753ec7e80afce58e"} err="failed to get container status \"99e4c3e77d4edddbe59c22e1f6be9b9efb3e296b5319db43753ec7e80afce58e\": rpc error: code = NotFound desc = could not find container \"99e4c3e77d4edddbe59c22e1f6be9b9efb3e296b5319db43753ec7e80afce58e\": container with ID starting with 99e4c3e77d4edddbe59c22e1f6be9b9efb3e296b5319db43753ec7e80afce58e not found: ID does not exist" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.892746 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:42:42 crc kubenswrapper[4744]: E1003 17:42:42.893081 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:42:42 crc kubenswrapper[4744]: I1003 17:42:42.909259 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8" path="/var/lib/kubelet/pods/a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8/volumes" Oct 03 17:42:50 crc kubenswrapper[4744]: I1003 17:42:50.852997 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7bf8c46bb6-7m9v6_068cb298-fa9c-4e22-92cc-0aa31310f185/barbican-api/0.log" Oct 03 17:42:50 crc kubenswrapper[4744]: I1003 17:42:50.896112 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7bf8c46bb6-7m9v6_068cb298-fa9c-4e22-92cc-0aa31310f185/barbican-api-log/0.log" Oct 03 17:42:51 crc kubenswrapper[4744]: I1003 17:42:51.090261 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6dfd6c4c58-rr7fr_6b4d5e00-5174-4dc3-b68e-185dedb764f1/barbican-keystone-listener/0.log" Oct 03 17:42:51 crc kubenswrapper[4744]: I1003 17:42:51.390132 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c46748695-jn5bb_9c69aee3-5a63-4011-b4ff-0afca3c16880/barbican-worker/0.log" Oct 03 17:42:51 crc kubenswrapper[4744]: I1003 17:42:51.707651 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c46748695-jn5bb_9c69aee3-5a63-4011-b4ff-0afca3c16880/barbican-worker-log/0.log" Oct 03 17:42:51 crc kubenswrapper[4744]: I1003 17:42:51.927404 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6dfd6c4c58-rr7fr_6b4d5e00-5174-4dc3-b68e-185dedb764f1/barbican-keystone-listener-log/0.log" Oct 03 17:42:51 crc kubenswrapper[4744]: I1003 17:42:51.976131 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9_b860d7da-e83d-4a61-aa1d-ba22cf817d8d/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:42:52 crc kubenswrapper[4744]: I1003 17:42:52.193703 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_13a29aaf-7761-4250-b9fd-bb69661f450e/ceilometer-notification-agent/0.log" Oct 03 17:42:52 crc kubenswrapper[4744]: I1003 17:42:52.201968 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_13a29aaf-7761-4250-b9fd-bb69661f450e/proxy-httpd/0.log" Oct 03 17:42:52 crc kubenswrapper[4744]: I1003 17:42:52.218473 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_13a29aaf-7761-4250-b9fd-bb69661f450e/ceilometer-central-agent/0.log" Oct 03 17:42:52 crc kubenswrapper[4744]: I1003 17:42:52.362852 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_13a29aaf-7761-4250-b9fd-bb69661f450e/sg-core/0.log" Oct 03 17:42:52 crc kubenswrapper[4744]: I1003 17:42:52.555011 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph_af690e2a-fac7-4b2b-9ca0-4473e9f6989d/ceph/0.log" Oct 03 17:42:52 crc kubenswrapper[4744]: I1003 17:42:52.841459 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_cee08598-c007-4049-8b75-d4be9c1b0796/cinder-api/0.log" Oct 03 17:42:52 crc kubenswrapper[4744]: I1003 17:42:52.930702 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_cee08598-c007-4049-8b75-d4be9c1b0796/cinder-api-log/0.log" Oct 03 17:42:53 crc kubenswrapper[4744]: I1003 17:42:53.194636 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_33874615-0e1f-4743-aa84-d262b6eab68c/probe/0.log" Oct 03 17:42:53 crc kubenswrapper[4744]: I1003 17:42:53.514546 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_008783bb-cc87-416a-9823-787ff756a3c9/probe/0.log" Oct 03 17:42:53 crc kubenswrapper[4744]: I1003 17:42:53.528724 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_008783bb-cc87-416a-9823-787ff756a3c9/cinder-scheduler/0.log" Oct 03 17:42:54 crc kubenswrapper[4744]: I1003 17:42:54.007581 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_146d5ce6-9a87-450e-b56c-371f2aa7af06/probe/0.log" Oct 03 17:42:54 crc kubenswrapper[4744]: I1003 17:42:54.281329 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg_4fe42e89-74c8-4ea1-b1f5-af28a032f3ce/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:42:54 crc kubenswrapper[4744]: I1003 17:42:54.493245 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c_2ac77289-8098-4cdd-892f-a56916c886e6/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:42:54 crc kubenswrapper[4744]: I1003 17:42:54.742267 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz_af14e991-2a54-4228-aa49-a8ce99d8eab2/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:42:54 crc kubenswrapper[4744]: I1003 17:42:54.898227 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:42:54 crc kubenswrapper[4744]: E1003 17:42:54.902953 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:42:54 crc kubenswrapper[4744]: I1003 17:42:54.978076 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5bb847fbb7-w6gpl_0bbdd34c-0174-4307-a779-0f8520f6e19a/init/0.log" Oct 03 17:42:55 crc kubenswrapper[4744]: I1003 17:42:55.181568 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5bb847fbb7-w6gpl_0bbdd34c-0174-4307-a779-0f8520f6e19a/init/0.log" Oct 03 17:42:55 crc kubenswrapper[4744]: I1003 17:42:55.596427 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5bb847fbb7-w6gpl_0bbdd34c-0174-4307-a779-0f8520f6e19a/dnsmasq-dns/0.log" Oct 03 17:42:55 crc kubenswrapper[4744]: I1003 17:42:55.618796 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8_b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:42:55 crc kubenswrapper[4744]: I1003 17:42:55.871115 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_411dd208-5058-4f51-b5ab-4e4f8c77cd8e/glance-httpd/0.log" Oct 03 17:42:56 crc kubenswrapper[4744]: I1003 17:42:56.065954 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_411dd208-5058-4f51-b5ab-4e4f8c77cd8e/glance-log/0.log" Oct 03 17:42:56 crc kubenswrapper[4744]: I1003 17:42:56.345913 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b7679ad4-70b7-4058-a2e6-c4dba8a409e9/glance-log/0.log" Oct 03 17:42:56 crc kubenswrapper[4744]: I1003 17:42:56.349972 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b7679ad4-70b7-4058-a2e6-c4dba8a409e9/glance-httpd/0.log" Oct 03 17:42:56 crc kubenswrapper[4744]: I1003 17:42:56.705825 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-85d8c4f7b-9pn28_39c711a0-3cea-4091-94a0-ec9df6e6a2d8/horizon/0.log" Oct 03 17:42:57 crc kubenswrapper[4744]: I1003 17:42:57.133070 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_146d5ce6-9a87-450e-b56c-371f2aa7af06/cinder-volume/0.log" Oct 03 17:42:57 crc kubenswrapper[4744]: I1003 17:42:57.202072 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc_3aa5f64b-0451-4034-822c-d462f01e7429/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:42:57 crc kubenswrapper[4744]: I1003 17:42:57.451469 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-jbg68_42a1ea96-7b56-4594-811d-7043e8640e39/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:42:57 crc kubenswrapper[4744]: I1003 17:42:57.504264 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-85d8c4f7b-9pn28_39c711a0-3cea-4091-94a0-ec9df6e6a2d8/horizon-log/0.log" Oct 03 17:42:57 crc kubenswrapper[4744]: I1003 17:42:57.863513 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29325181-pc2dm_513c18a1-42b9-43f1-bb0c-4a99a3c45ea6/keystone-cron/0.log" Oct 03 17:42:58 crc kubenswrapper[4744]: I1003 17:42:58.085147 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_6c69a90c-bccf-4209-94e1-fe80d4914377/kube-state-metrics/0.log" Oct 03 17:42:58 crc kubenswrapper[4744]: I1003 17:42:58.120818 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_33874615-0e1f-4743-aa84-d262b6eab68c/cinder-backup/0.log" Oct 03 17:42:58 crc kubenswrapper[4744]: I1003 17:42:58.356516 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5_78b5d479-1346-45e4-96f9-310b6b32ec50/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:42:58 crc kubenswrapper[4744]: I1003 17:42:58.963173 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_f392aa4a-226f-4009-acbb-d66b4aaeb95e/manila-api/0.log" Oct 03 17:42:59 crc kubenswrapper[4744]: I1003 17:42:59.137389 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_75efca1b-e080-4620-ad1b-8e1bcc3ba3ba/manila-scheduler/0.log" Oct 03 17:42:59 crc kubenswrapper[4744]: I1003 17:42:59.139342 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_75efca1b-e080-4620-ad1b-8e1bcc3ba3ba/probe/0.log" Oct 03 17:42:59 crc kubenswrapper[4744]: I1003 17:42:59.571906 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_3c066c82-d9c3-40bc-805d-16a45e328e49/probe/0.log" Oct 03 17:42:59 crc kubenswrapper[4744]: I1003 17:42:59.626822 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_f392aa4a-226f-4009-acbb-d66b4aaeb95e/manila-api-log/0.log" Oct 03 17:42:59 crc kubenswrapper[4744]: I1003 17:42:59.797746 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_3c066c82-d9c3-40bc-805d-16a45e328e49/manila-share/0.log" Oct 03 17:43:01 crc kubenswrapper[4744]: I1003 17:43:01.307806 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6bc59588f-s74gr_ec5a5a4c-4896-4629-be5d-2b96f2729eaf/neutron-httpd/0.log" Oct 03 17:43:01 crc kubenswrapper[4744]: I1003 17:43:01.416335 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6589b6d4db-rbbtp_6e8f7c20-0249-41e8-9cc6-915cf0762b3a/keystone-api/0.log" Oct 03 17:43:01 crc kubenswrapper[4744]: I1003 17:43:01.493593 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6bc59588f-s74gr_ec5a5a4c-4896-4629-be5d-2b96f2729eaf/neutron-api/0.log" Oct 03 17:43:01 crc kubenswrapper[4744]: I1003 17:43:01.680697 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz_a1fb9143-c0e9-44b2-b80d-ff1ed760a115/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:43:03 crc kubenswrapper[4744]: I1003 17:43:03.152801 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_da322563-8083-4e70-b339-a0c68c23e0c3/nova-cell0-conductor-conductor/0.log" Oct 03 17:43:03 crc kubenswrapper[4744]: I1003 17:43:03.785637 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_74a40f5d-b401-437e-9e4c-38e1d3459e40/nova-cell1-conductor-conductor/0.log" Oct 03 17:43:03 crc kubenswrapper[4744]: I1003 17:43:03.799566 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92/nova-api-log/0.log" Oct 03 17:43:04 crc kubenswrapper[4744]: I1003 17:43:04.136327 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_8005560a-4d21-4e19-9e47-17a703457765/nova-cell1-novncproxy-novncproxy/0.log" Oct 03 17:43:04 crc kubenswrapper[4744]: I1003 17:43:04.402327 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92/nova-api-api/0.log" Oct 03 17:43:04 crc kubenswrapper[4744]: I1003 17:43:04.432720 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-z279b_710d1522-5c50-4e85-bb48-a099cfbb892d/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:43:04 crc kubenswrapper[4744]: I1003 17:43:04.729976 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_76b30dc4-9f12-483e-874e-6654e45e521a/nova-metadata-log/0.log" Oct 03 17:43:05 crc kubenswrapper[4744]: I1003 17:43:05.239757 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8/mysql-bootstrap/0.log" Oct 03 17:43:05 crc kubenswrapper[4744]: I1003 17:43:05.378171 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1ae57486-d534-4c67-a235-218fe8293726/nova-scheduler-scheduler/0.log" Oct 03 17:43:05 crc kubenswrapper[4744]: I1003 17:43:05.476911 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8/mysql-bootstrap/0.log" Oct 03 17:43:05 crc kubenswrapper[4744]: I1003 17:43:05.608700 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8/galera/0.log" Oct 03 17:43:05 crc kubenswrapper[4744]: I1003 17:43:05.839242 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7c8031bb-6ec3-42b3-ace7-80e25071e230/mysql-bootstrap/0.log" Oct 03 17:43:06 crc kubenswrapper[4744]: I1003 17:43:06.058105 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7c8031bb-6ec3-42b3-ace7-80e25071e230/mysql-bootstrap/0.log" Oct 03 17:43:06 crc kubenswrapper[4744]: I1003 17:43:06.088334 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7c8031bb-6ec3-42b3-ace7-80e25071e230/galera/0.log" Oct 03 17:43:06 crc kubenswrapper[4744]: I1003 17:43:06.276197 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1/openstackclient/0.log" Oct 03 17:43:06 crc kubenswrapper[4744]: I1003 17:43:06.331350 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_76b30dc4-9f12-483e-874e-6654e45e521a/nova-metadata-metadata/0.log" Oct 03 17:43:06 crc kubenswrapper[4744]: I1003 17:43:06.505290 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4jn8x_bc8fb837-d1e5-4953-ae6c-fbee5f867e49/ovn-controller/0.log" Oct 03 17:43:06 crc kubenswrapper[4744]: I1003 17:43:06.727572 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-mgnmg_02bac763-b977-4d19-a6fd-300b0b2cefc3/openstack-network-exporter/0.log" Oct 03 17:43:06 crc kubenswrapper[4744]: I1003 17:43:06.857709 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-99k4w_77d29fac-2896-4f1a-a193-f4b559f1e9b6/ovsdb-server-init/0.log" Oct 03 17:43:06 crc kubenswrapper[4744]: I1003 17:43:06.892179 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:43:06 crc kubenswrapper[4744]: E1003 17:43:06.892441 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:43:06 crc kubenswrapper[4744]: I1003 17:43:06.995135 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-99k4w_77d29fac-2896-4f1a-a193-f4b559f1e9b6/ovsdb-server-init/0.log" Oct 03 17:43:07 crc kubenswrapper[4744]: I1003 17:43:07.070701 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-99k4w_77d29fac-2896-4f1a-a193-f4b559f1e9b6/ovsdb-server/0.log" Oct 03 17:43:07 crc kubenswrapper[4744]: I1003 17:43:07.086552 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-99k4w_77d29fac-2896-4f1a-a193-f4b559f1e9b6/ovs-vswitchd/0.log" Oct 03 17:43:07 crc kubenswrapper[4744]: I1003 17:43:07.276913 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-8x7d2_6c864721-c102-4044-bd81-74f3fbe29040/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:43:07 crc kubenswrapper[4744]: I1003 17:43:07.445243 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_09af7ba6-34b5-4454-9dae-5bcc301ed20c/openstack-network-exporter/0.log" Oct 03 17:43:07 crc kubenswrapper[4744]: I1003 17:43:07.502039 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_09af7ba6-34b5-4454-9dae-5bcc301ed20c/ovn-northd/0.log" Oct 03 17:43:07 crc kubenswrapper[4744]: I1003 17:43:07.677416 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_803d0f2b-9d7d-4d98-909c-18ffdec34aed/openstack-network-exporter/0.log" Oct 03 17:43:07 crc kubenswrapper[4744]: I1003 17:43:07.696613 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_803d0f2b-9d7d-4d98-909c-18ffdec34aed/ovsdbserver-nb/0.log" Oct 03 17:43:07 crc kubenswrapper[4744]: I1003 17:43:07.885886 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7c9e2064-beda-46d7-9221-06b2bbb375eb/openstack-network-exporter/0.log" Oct 03 17:43:07 crc kubenswrapper[4744]: I1003 17:43:07.891249 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7c9e2064-beda-46d7-9221-06b2bbb375eb/ovsdbserver-sb/0.log" Oct 03 17:43:08 crc kubenswrapper[4744]: I1003 17:43:08.479110 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-774fb64f68-qgzkk_0d2b9c37-6d79-420d-bc4e-3d5888b5d61e/placement-api/0.log" Oct 03 17:43:08 crc kubenswrapper[4744]: I1003 17:43:08.586224 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-774fb64f68-qgzkk_0d2b9c37-6d79-420d-bc4e-3d5888b5d61e/placement-log/0.log" Oct 03 17:43:08 crc kubenswrapper[4744]: I1003 17:43:08.628840 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d34a40f-b279-4d50-88c9-bb7776fc678b/setup-container/0.log" Oct 03 17:43:08 crc kubenswrapper[4744]: I1003 17:43:08.761542 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d34a40f-b279-4d50-88c9-bb7776fc678b/setup-container/0.log" Oct 03 17:43:08 crc kubenswrapper[4744]: I1003 17:43:08.828749 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d34a40f-b279-4d50-88c9-bb7776fc678b/rabbitmq/0.log" Oct 03 17:43:08 crc kubenswrapper[4744]: I1003 17:43:08.959833 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_52159513-8a05-44c2-89f2-fb6d08e50b36/setup-container/0.log" Oct 03 17:43:09 crc kubenswrapper[4744]: I1003 17:43:09.184799 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_52159513-8a05-44c2-89f2-fb6d08e50b36/setup-container/0.log" Oct 03 17:43:09 crc kubenswrapper[4744]: I1003 17:43:09.251752 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_52159513-8a05-44c2-89f2-fb6d08e50b36/rabbitmq/0.log" Oct 03 17:43:09 crc kubenswrapper[4744]: I1003 17:43:09.445860 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc_0f862434-d64e-4c02-b8de-a120185d0053/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:43:09 crc kubenswrapper[4744]: I1003 17:43:09.491066 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-tr7p9_9f6a23c2-45cf-4443-b1ef-befcea950ae8/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:43:09 crc kubenswrapper[4744]: I1003 17:43:09.732295 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7_962e48ea-a515-44c9-b2c2-2e9705a31455/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:43:09 crc kubenswrapper[4744]: I1003 17:43:09.941135 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-mf5dj_1b94f190-a56d-420d-9558-7ec1116b90b9/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:43:10 crc kubenswrapper[4744]: I1003 17:43:10.011441 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8nbg9_1700fe09-1277-4cda-89dc-03d220941b9a/ssh-known-hosts-edpm-deployment/0.log" Oct 03 17:43:10 crc kubenswrapper[4744]: I1003 17:43:10.235087 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-689c67db59-x766r_1c64dd03-9cb9-402f-9697-a23538c9ebc5/proxy-server/0.log" Oct 03 17:43:10 crc kubenswrapper[4744]: I1003 17:43:10.378070 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-689c67db59-x766r_1c64dd03-9cb9-402f-9697-a23538c9ebc5/proxy-httpd/0.log" Oct 03 17:43:10 crc kubenswrapper[4744]: I1003 17:43:10.437226 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-dsc89_0b88692d-9c4d-4c67-b70f-921810ca71e1/swift-ring-rebalance/0.log" Oct 03 17:43:10 crc kubenswrapper[4744]: I1003 17:43:10.721373 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/account-auditor/0.log" Oct 03 17:43:10 crc kubenswrapper[4744]: I1003 17:43:10.728219 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/account-reaper/0.log" Oct 03 17:43:10 crc kubenswrapper[4744]: I1003 17:43:10.859749 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/account-replicator/0.log" Oct 03 17:43:10 crc kubenswrapper[4744]: I1003 17:43:10.949776 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/account-server/0.log" Oct 03 17:43:10 crc kubenswrapper[4744]: I1003 17:43:10.987691 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/container-auditor/0.log" Oct 03 17:43:11 crc kubenswrapper[4744]: I1003 17:43:11.065829 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7/memcached/0.log" Oct 03 17:43:11 crc kubenswrapper[4744]: I1003 17:43:11.092886 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/container-replicator/0.log" Oct 03 17:43:11 crc kubenswrapper[4744]: I1003 17:43:11.140807 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/container-server/0.log" Oct 03 17:43:11 crc kubenswrapper[4744]: I1003 17:43:11.186153 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/container-updater/0.log" Oct 03 17:43:11 crc kubenswrapper[4744]: I1003 17:43:11.247521 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/object-auditor/0.log" Oct 03 17:43:11 crc kubenswrapper[4744]: I1003 17:43:11.273516 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/object-expirer/0.log" Oct 03 17:43:11 crc kubenswrapper[4744]: I1003 17:43:11.951782 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/object-replicator/0.log" Oct 03 17:43:11 crc kubenswrapper[4744]: I1003 17:43:11.954656 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/object-server/0.log" Oct 03 17:43:11 crc kubenswrapper[4744]: I1003 17:43:11.961900 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/object-updater/0.log" Oct 03 17:43:11 crc kubenswrapper[4744]: I1003 17:43:11.973002 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/rsync/0.log" Oct 03 17:43:12 crc kubenswrapper[4744]: I1003 17:43:12.172574 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/swift-recon-cron/0.log" Oct 03 17:43:12 crc kubenswrapper[4744]: I1003 17:43:12.256083 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn_4d4a0574-77fd-40fe-aa97-c4a26255473e/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:43:12 crc kubenswrapper[4744]: I1003 17:43:12.404389 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_812d60e3-44e3-4f22-9afa-448f0ac216ef/test-operator-logs-container/0.log" Oct 03 17:43:12 crc kubenswrapper[4744]: I1003 17:43:12.410761 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_595f783e-db00-49b2-968f-79163001ffdd/tempest-tests-tempest-tests-runner/0.log" Oct 03 17:43:12 crc kubenswrapper[4744]: I1003 17:43:12.569132 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7_d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:43:21 crc kubenswrapper[4744]: I1003 17:43:21.892381 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:43:21 crc kubenswrapper[4744]: E1003 17:43:21.893610 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.160429 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5tdvw"] Oct 03 17:43:27 crc kubenswrapper[4744]: E1003 17:43:27.161283 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8" containerName="extract-utilities" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.161300 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8" containerName="extract-utilities" Oct 03 17:43:27 crc kubenswrapper[4744]: E1003 17:43:27.161324 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8" containerName="extract-content" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.161330 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8" containerName="extract-content" Oct 03 17:43:27 crc kubenswrapper[4744]: E1003 17:43:27.161354 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8" containerName="registry-server" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.161361 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8" containerName="registry-server" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.165673 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7e8d457-eb3f-4aad-9ee8-73abdd7cd3e8" containerName="registry-server" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.167712 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.179765 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5tdvw"] Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.237047 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275e600b-eda6-4a3a-b044-e7a014a4878d-catalog-content\") pod \"redhat-marketplace-5tdvw\" (UID: \"275e600b-eda6-4a3a-b044-e7a014a4878d\") " pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.237135 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dwtn\" (UniqueName: \"kubernetes.io/projected/275e600b-eda6-4a3a-b044-e7a014a4878d-kube-api-access-8dwtn\") pod \"redhat-marketplace-5tdvw\" (UID: \"275e600b-eda6-4a3a-b044-e7a014a4878d\") " pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.237205 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275e600b-eda6-4a3a-b044-e7a014a4878d-utilities\") pod \"redhat-marketplace-5tdvw\" (UID: \"275e600b-eda6-4a3a-b044-e7a014a4878d\") " pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.339405 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275e600b-eda6-4a3a-b044-e7a014a4878d-catalog-content\") pod \"redhat-marketplace-5tdvw\" (UID: \"275e600b-eda6-4a3a-b044-e7a014a4878d\") " pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.339515 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dwtn\" (UniqueName: \"kubernetes.io/projected/275e600b-eda6-4a3a-b044-e7a014a4878d-kube-api-access-8dwtn\") pod \"redhat-marketplace-5tdvw\" (UID: \"275e600b-eda6-4a3a-b044-e7a014a4878d\") " pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.339594 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275e600b-eda6-4a3a-b044-e7a014a4878d-utilities\") pod \"redhat-marketplace-5tdvw\" (UID: \"275e600b-eda6-4a3a-b044-e7a014a4878d\") " pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.340031 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275e600b-eda6-4a3a-b044-e7a014a4878d-utilities\") pod \"redhat-marketplace-5tdvw\" (UID: \"275e600b-eda6-4a3a-b044-e7a014a4878d\") " pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.340146 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275e600b-eda6-4a3a-b044-e7a014a4878d-catalog-content\") pod \"redhat-marketplace-5tdvw\" (UID: \"275e600b-eda6-4a3a-b044-e7a014a4878d\") " pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.356854 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dwtn\" (UniqueName: \"kubernetes.io/projected/275e600b-eda6-4a3a-b044-e7a014a4878d-kube-api-access-8dwtn\") pod \"redhat-marketplace-5tdvw\" (UID: \"275e600b-eda6-4a3a-b044-e7a014a4878d\") " pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:27 crc kubenswrapper[4744]: I1003 17:43:27.501932 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:28 crc kubenswrapper[4744]: I1003 17:43:28.187419 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5tdvw"] Oct 03 17:43:29 crc kubenswrapper[4744]: I1003 17:43:29.061313 4744 generic.go:334] "Generic (PLEG): container finished" podID="275e600b-eda6-4a3a-b044-e7a014a4878d" containerID="a57d7309b63399ddf58132502b57ce9ac48ca55989fb7b30a3c059b54c863bed" exitCode=0 Oct 03 17:43:29 crc kubenswrapper[4744]: I1003 17:43:29.061367 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tdvw" event={"ID":"275e600b-eda6-4a3a-b044-e7a014a4878d","Type":"ContainerDied","Data":"a57d7309b63399ddf58132502b57ce9ac48ca55989fb7b30a3c059b54c863bed"} Oct 03 17:43:29 crc kubenswrapper[4744]: I1003 17:43:29.061964 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tdvw" event={"ID":"275e600b-eda6-4a3a-b044-e7a014a4878d","Type":"ContainerStarted","Data":"2039d2dd54540c913050773e79a0db1a206b4ebb73184e530b4ceb5e5165c1a3"} Oct 03 17:43:31 crc kubenswrapper[4744]: I1003 17:43:31.083700 4744 generic.go:334] "Generic (PLEG): container finished" podID="275e600b-eda6-4a3a-b044-e7a014a4878d" containerID="1b41b9ab387e46537f8c484a07834cab12aeecfeda1339bd1804e8a741c8ed65" exitCode=0 Oct 03 17:43:31 crc kubenswrapper[4744]: I1003 17:43:31.083920 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tdvw" event={"ID":"275e600b-eda6-4a3a-b044-e7a014a4878d","Type":"ContainerDied","Data":"1b41b9ab387e46537f8c484a07834cab12aeecfeda1339bd1804e8a741c8ed65"} Oct 03 17:43:32 crc kubenswrapper[4744]: I1003 17:43:32.093943 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tdvw" event={"ID":"275e600b-eda6-4a3a-b044-e7a014a4878d","Type":"ContainerStarted","Data":"dd74e22b0df1e305d49be49f62cf369516e0f64a679bd4f04b3a0b5db0386f59"} Oct 03 17:43:32 crc kubenswrapper[4744]: I1003 17:43:32.125049 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5tdvw" podStartSLOduration=2.6887925409999998 podStartE2EDuration="5.125028752s" podCreationTimestamp="2025-10-03 17:43:27 +0000 UTC" firstStartedPulling="2025-10-03 17:43:29.063179824 +0000 UTC m=+4735.343055720" lastFinishedPulling="2025-10-03 17:43:31.499415995 +0000 UTC m=+4737.779291931" observedRunningTime="2025-10-03 17:43:32.117265128 +0000 UTC m=+4738.397141024" watchObservedRunningTime="2025-10-03 17:43:32.125028752 +0000 UTC m=+4738.404904648" Oct 03 17:43:33 crc kubenswrapper[4744]: I1003 17:43:33.892412 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:43:33 crc kubenswrapper[4744]: E1003 17:43:33.892945 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:43:37 crc kubenswrapper[4744]: I1003 17:43:37.502355 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:37 crc kubenswrapper[4744]: I1003 17:43:37.502960 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:37 crc kubenswrapper[4744]: I1003 17:43:37.554552 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:38 crc kubenswrapper[4744]: I1003 17:43:38.203473 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:38 crc kubenswrapper[4744]: I1003 17:43:38.258978 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5tdvw"] Oct 03 17:43:40 crc kubenswrapper[4744]: I1003 17:43:40.175704 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5tdvw" podUID="275e600b-eda6-4a3a-b044-e7a014a4878d" containerName="registry-server" containerID="cri-o://dd74e22b0df1e305d49be49f62cf369516e0f64a679bd4f04b3a0b5db0386f59" gracePeriod=2 Oct 03 17:43:41 crc kubenswrapper[4744]: I1003 17:43:41.186899 4744 generic.go:334] "Generic (PLEG): container finished" podID="275e600b-eda6-4a3a-b044-e7a014a4878d" containerID="dd74e22b0df1e305d49be49f62cf369516e0f64a679bd4f04b3a0b5db0386f59" exitCode=0 Oct 03 17:43:41 crc kubenswrapper[4744]: I1003 17:43:41.186975 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tdvw" event={"ID":"275e600b-eda6-4a3a-b044-e7a014a4878d","Type":"ContainerDied","Data":"dd74e22b0df1e305d49be49f62cf369516e0f64a679bd4f04b3a0b5db0386f59"} Oct 03 17:43:41 crc kubenswrapper[4744]: I1003 17:43:41.784353 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:41 crc kubenswrapper[4744]: I1003 17:43:41.942725 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275e600b-eda6-4a3a-b044-e7a014a4878d-utilities\") pod \"275e600b-eda6-4a3a-b044-e7a014a4878d\" (UID: \"275e600b-eda6-4a3a-b044-e7a014a4878d\") " Oct 03 17:43:41 crc kubenswrapper[4744]: I1003 17:43:41.942776 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275e600b-eda6-4a3a-b044-e7a014a4878d-catalog-content\") pod \"275e600b-eda6-4a3a-b044-e7a014a4878d\" (UID: \"275e600b-eda6-4a3a-b044-e7a014a4878d\") " Oct 03 17:43:41 crc kubenswrapper[4744]: I1003 17:43:41.942959 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dwtn\" (UniqueName: \"kubernetes.io/projected/275e600b-eda6-4a3a-b044-e7a014a4878d-kube-api-access-8dwtn\") pod \"275e600b-eda6-4a3a-b044-e7a014a4878d\" (UID: \"275e600b-eda6-4a3a-b044-e7a014a4878d\") " Oct 03 17:43:41 crc kubenswrapper[4744]: I1003 17:43:41.943713 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/275e600b-eda6-4a3a-b044-e7a014a4878d-utilities" (OuterVolumeSpecName: "utilities") pod "275e600b-eda6-4a3a-b044-e7a014a4878d" (UID: "275e600b-eda6-4a3a-b044-e7a014a4878d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:43:41 crc kubenswrapper[4744]: I1003 17:43:41.950474 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/275e600b-eda6-4a3a-b044-e7a014a4878d-kube-api-access-8dwtn" (OuterVolumeSpecName: "kube-api-access-8dwtn") pod "275e600b-eda6-4a3a-b044-e7a014a4878d" (UID: "275e600b-eda6-4a3a-b044-e7a014a4878d"). InnerVolumeSpecName "kube-api-access-8dwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:43:41 crc kubenswrapper[4744]: I1003 17:43:41.959031 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/275e600b-eda6-4a3a-b044-e7a014a4878d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "275e600b-eda6-4a3a-b044-e7a014a4878d" (UID: "275e600b-eda6-4a3a-b044-e7a014a4878d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:43:42 crc kubenswrapper[4744]: I1003 17:43:42.047444 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dwtn\" (UniqueName: \"kubernetes.io/projected/275e600b-eda6-4a3a-b044-e7a014a4878d-kube-api-access-8dwtn\") on node \"crc\" DevicePath \"\"" Oct 03 17:43:42 crc kubenswrapper[4744]: I1003 17:43:42.047476 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/275e600b-eda6-4a3a-b044-e7a014a4878d-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:43:42 crc kubenswrapper[4744]: I1003 17:43:42.047485 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/275e600b-eda6-4a3a-b044-e7a014a4878d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:43:42 crc kubenswrapper[4744]: I1003 17:43:42.199073 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5tdvw" event={"ID":"275e600b-eda6-4a3a-b044-e7a014a4878d","Type":"ContainerDied","Data":"2039d2dd54540c913050773e79a0db1a206b4ebb73184e530b4ceb5e5165c1a3"} Oct 03 17:43:42 crc kubenswrapper[4744]: I1003 17:43:42.199121 4744 scope.go:117] "RemoveContainer" containerID="dd74e22b0df1e305d49be49f62cf369516e0f64a679bd4f04b3a0b5db0386f59" Oct 03 17:43:42 crc kubenswrapper[4744]: I1003 17:43:42.199226 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5tdvw" Oct 03 17:43:42 crc kubenswrapper[4744]: I1003 17:43:42.231906 4744 scope.go:117] "RemoveContainer" containerID="1b41b9ab387e46537f8c484a07834cab12aeecfeda1339bd1804e8a741c8ed65" Oct 03 17:43:42 crc kubenswrapper[4744]: I1003 17:43:42.249683 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5tdvw"] Oct 03 17:43:42 crc kubenswrapper[4744]: I1003 17:43:42.260447 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5tdvw"] Oct 03 17:43:42 crc kubenswrapper[4744]: I1003 17:43:42.544148 4744 scope.go:117] "RemoveContainer" containerID="a57d7309b63399ddf58132502b57ce9ac48ca55989fb7b30a3c059b54c863bed" Oct 03 17:43:42 crc kubenswrapper[4744]: I1003 17:43:42.905713 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="275e600b-eda6-4a3a-b044-e7a014a4878d" path="/var/lib/kubelet/pods/275e600b-eda6-4a3a-b044-e7a014a4878d/volumes" Oct 03 17:43:46 crc kubenswrapper[4744]: I1003 17:43:46.892424 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:43:46 crc kubenswrapper[4744]: E1003 17:43:46.893809 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:43:57 crc kubenswrapper[4744]: I1003 17:43:57.368939 4744 generic.go:334] "Generic (PLEG): container finished" podID="292f592c-6a56-48eb-acda-07416514c6ab" containerID="01b8eec269bfadcd5704da9f37361eaa3d2894284c41385246ea3e7f2e6dd0a6" exitCode=0 Oct 03 17:43:57 crc kubenswrapper[4744]: I1003 17:43:57.369050 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mhn28/crc-debug-6hv6g" event={"ID":"292f592c-6a56-48eb-acda-07416514c6ab","Type":"ContainerDied","Data":"01b8eec269bfadcd5704da9f37361eaa3d2894284c41385246ea3e7f2e6dd0a6"} Oct 03 17:43:58 crc kubenswrapper[4744]: I1003 17:43:58.517694 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/crc-debug-6hv6g" Oct 03 17:43:58 crc kubenswrapper[4744]: I1003 17:43:58.560304 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mhn28/crc-debug-6hv6g"] Oct 03 17:43:58 crc kubenswrapper[4744]: I1003 17:43:58.570575 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mhn28/crc-debug-6hv6g"] Oct 03 17:43:58 crc kubenswrapper[4744]: I1003 17:43:58.608182 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/292f592c-6a56-48eb-acda-07416514c6ab-host\") pod \"292f592c-6a56-48eb-acda-07416514c6ab\" (UID: \"292f592c-6a56-48eb-acda-07416514c6ab\") " Oct 03 17:43:58 crc kubenswrapper[4744]: I1003 17:43:58.608331 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/292f592c-6a56-48eb-acda-07416514c6ab-host" (OuterVolumeSpecName: "host") pod "292f592c-6a56-48eb-acda-07416514c6ab" (UID: "292f592c-6a56-48eb-acda-07416514c6ab"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 17:43:58 crc kubenswrapper[4744]: I1003 17:43:58.608418 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl4k2\" (UniqueName: \"kubernetes.io/projected/292f592c-6a56-48eb-acda-07416514c6ab-kube-api-access-kl4k2\") pod \"292f592c-6a56-48eb-acda-07416514c6ab\" (UID: \"292f592c-6a56-48eb-acda-07416514c6ab\") " Oct 03 17:43:58 crc kubenswrapper[4744]: I1003 17:43:58.609147 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/292f592c-6a56-48eb-acda-07416514c6ab-host\") on node \"crc\" DevicePath \"\"" Oct 03 17:43:58 crc kubenswrapper[4744]: I1003 17:43:58.616983 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/292f592c-6a56-48eb-acda-07416514c6ab-kube-api-access-kl4k2" (OuterVolumeSpecName: "kube-api-access-kl4k2") pod "292f592c-6a56-48eb-acda-07416514c6ab" (UID: "292f592c-6a56-48eb-acda-07416514c6ab"). InnerVolumeSpecName "kube-api-access-kl4k2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:43:58 crc kubenswrapper[4744]: I1003 17:43:58.711176 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl4k2\" (UniqueName: \"kubernetes.io/projected/292f592c-6a56-48eb-acda-07416514c6ab-kube-api-access-kl4k2\") on node \"crc\" DevicePath \"\"" Oct 03 17:43:58 crc kubenswrapper[4744]: I1003 17:43:58.892218 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:43:58 crc kubenswrapper[4744]: E1003 17:43:58.892897 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:43:58 crc kubenswrapper[4744]: I1003 17:43:58.904317 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="292f592c-6a56-48eb-acda-07416514c6ab" path="/var/lib/kubelet/pods/292f592c-6a56-48eb-acda-07416514c6ab/volumes" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.403154 4744 scope.go:117] "RemoveContainer" containerID="01b8eec269bfadcd5704da9f37361eaa3d2894284c41385246ea3e7f2e6dd0a6" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.403381 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/crc-debug-6hv6g" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.778335 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mhn28/crc-debug-wq9vc"] Oct 03 17:43:59 crc kubenswrapper[4744]: E1003 17:43:59.779354 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="275e600b-eda6-4a3a-b044-e7a014a4878d" containerName="extract-utilities" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.779378 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="275e600b-eda6-4a3a-b044-e7a014a4878d" containerName="extract-utilities" Oct 03 17:43:59 crc kubenswrapper[4744]: E1003 17:43:59.779399 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="275e600b-eda6-4a3a-b044-e7a014a4878d" containerName="extract-content" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.779411 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="275e600b-eda6-4a3a-b044-e7a014a4878d" containerName="extract-content" Oct 03 17:43:59 crc kubenswrapper[4744]: E1003 17:43:59.779457 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="275e600b-eda6-4a3a-b044-e7a014a4878d" containerName="registry-server" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.779467 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="275e600b-eda6-4a3a-b044-e7a014a4878d" containerName="registry-server" Oct 03 17:43:59 crc kubenswrapper[4744]: E1003 17:43:59.779482 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="292f592c-6a56-48eb-acda-07416514c6ab" containerName="container-00" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.779514 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="292f592c-6a56-48eb-acda-07416514c6ab" containerName="container-00" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.780404 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="275e600b-eda6-4a3a-b044-e7a014a4878d" containerName="registry-server" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.780534 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="292f592c-6a56-48eb-acda-07416514c6ab" containerName="container-00" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.781440 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/crc-debug-wq9vc" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.838393 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzvc8\" (UniqueName: \"kubernetes.io/projected/c963dd0b-c455-4f0c-8aa5-c3f08cc3d644-kube-api-access-jzvc8\") pod \"crc-debug-wq9vc\" (UID: \"c963dd0b-c455-4f0c-8aa5-c3f08cc3d644\") " pod="openshift-must-gather-mhn28/crc-debug-wq9vc" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.838688 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c963dd0b-c455-4f0c-8aa5-c3f08cc3d644-host\") pod \"crc-debug-wq9vc\" (UID: \"c963dd0b-c455-4f0c-8aa5-c3f08cc3d644\") " pod="openshift-must-gather-mhn28/crc-debug-wq9vc" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.939948 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c963dd0b-c455-4f0c-8aa5-c3f08cc3d644-host\") pod \"crc-debug-wq9vc\" (UID: \"c963dd0b-c455-4f0c-8aa5-c3f08cc3d644\") " pod="openshift-must-gather-mhn28/crc-debug-wq9vc" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.940064 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzvc8\" (UniqueName: \"kubernetes.io/projected/c963dd0b-c455-4f0c-8aa5-c3f08cc3d644-kube-api-access-jzvc8\") pod \"crc-debug-wq9vc\" (UID: \"c963dd0b-c455-4f0c-8aa5-c3f08cc3d644\") " pod="openshift-must-gather-mhn28/crc-debug-wq9vc" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.940061 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c963dd0b-c455-4f0c-8aa5-c3f08cc3d644-host\") pod \"crc-debug-wq9vc\" (UID: \"c963dd0b-c455-4f0c-8aa5-c3f08cc3d644\") " pod="openshift-must-gather-mhn28/crc-debug-wq9vc" Oct 03 17:43:59 crc kubenswrapper[4744]: I1003 17:43:59.962896 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzvc8\" (UniqueName: \"kubernetes.io/projected/c963dd0b-c455-4f0c-8aa5-c3f08cc3d644-kube-api-access-jzvc8\") pod \"crc-debug-wq9vc\" (UID: \"c963dd0b-c455-4f0c-8aa5-c3f08cc3d644\") " pod="openshift-must-gather-mhn28/crc-debug-wq9vc" Oct 03 17:44:00 crc kubenswrapper[4744]: I1003 17:44:00.104305 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/crc-debug-wq9vc" Oct 03 17:44:00 crc kubenswrapper[4744]: I1003 17:44:00.416968 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mhn28/crc-debug-wq9vc" event={"ID":"c963dd0b-c455-4f0c-8aa5-c3f08cc3d644","Type":"ContainerStarted","Data":"058e5cc3ba56f7d11383f600b193dd1611112103fea403dc287147f26667428c"} Oct 03 17:44:00 crc kubenswrapper[4744]: I1003 17:44:00.417334 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mhn28/crc-debug-wq9vc" event={"ID":"c963dd0b-c455-4f0c-8aa5-c3f08cc3d644","Type":"ContainerStarted","Data":"4d46e461807abb65fddb75322c7678119c73ed3c2fa334c701ff11572be6cba8"} Oct 03 17:44:00 crc kubenswrapper[4744]: I1003 17:44:00.444627 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mhn28/crc-debug-wq9vc" podStartSLOduration=1.444611726 podStartE2EDuration="1.444611726s" podCreationTimestamp="2025-10-03 17:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 17:44:00.432472633 +0000 UTC m=+4766.712348529" watchObservedRunningTime="2025-10-03 17:44:00.444611726 +0000 UTC m=+4766.724487622" Oct 03 17:44:01 crc kubenswrapper[4744]: I1003 17:44:01.429658 4744 generic.go:334] "Generic (PLEG): container finished" podID="c963dd0b-c455-4f0c-8aa5-c3f08cc3d644" containerID="058e5cc3ba56f7d11383f600b193dd1611112103fea403dc287147f26667428c" exitCode=0 Oct 03 17:44:01 crc kubenswrapper[4744]: I1003 17:44:01.429724 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mhn28/crc-debug-wq9vc" event={"ID":"c963dd0b-c455-4f0c-8aa5-c3f08cc3d644","Type":"ContainerDied","Data":"058e5cc3ba56f7d11383f600b193dd1611112103fea403dc287147f26667428c"} Oct 03 17:44:03 crc kubenswrapper[4744]: I1003 17:44:03.224463 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/crc-debug-wq9vc" Oct 03 17:44:03 crc kubenswrapper[4744]: I1003 17:44:03.393060 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzvc8\" (UniqueName: \"kubernetes.io/projected/c963dd0b-c455-4f0c-8aa5-c3f08cc3d644-kube-api-access-jzvc8\") pod \"c963dd0b-c455-4f0c-8aa5-c3f08cc3d644\" (UID: \"c963dd0b-c455-4f0c-8aa5-c3f08cc3d644\") " Oct 03 17:44:03 crc kubenswrapper[4744]: I1003 17:44:03.393180 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c963dd0b-c455-4f0c-8aa5-c3f08cc3d644-host\") pod \"c963dd0b-c455-4f0c-8aa5-c3f08cc3d644\" (UID: \"c963dd0b-c455-4f0c-8aa5-c3f08cc3d644\") " Oct 03 17:44:03 crc kubenswrapper[4744]: I1003 17:44:03.393888 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c963dd0b-c455-4f0c-8aa5-c3f08cc3d644-host" (OuterVolumeSpecName: "host") pod "c963dd0b-c455-4f0c-8aa5-c3f08cc3d644" (UID: "c963dd0b-c455-4f0c-8aa5-c3f08cc3d644"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 17:44:03 crc kubenswrapper[4744]: I1003 17:44:03.410113 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c963dd0b-c455-4f0c-8aa5-c3f08cc3d644-kube-api-access-jzvc8" (OuterVolumeSpecName: "kube-api-access-jzvc8") pod "c963dd0b-c455-4f0c-8aa5-c3f08cc3d644" (UID: "c963dd0b-c455-4f0c-8aa5-c3f08cc3d644"). InnerVolumeSpecName "kube-api-access-jzvc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:44:03 crc kubenswrapper[4744]: I1003 17:44:03.453380 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mhn28/crc-debug-wq9vc" event={"ID":"c963dd0b-c455-4f0c-8aa5-c3f08cc3d644","Type":"ContainerDied","Data":"4d46e461807abb65fddb75322c7678119c73ed3c2fa334c701ff11572be6cba8"} Oct 03 17:44:03 crc kubenswrapper[4744]: I1003 17:44:03.453643 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d46e461807abb65fddb75322c7678119c73ed3c2fa334c701ff11572be6cba8" Oct 03 17:44:03 crc kubenswrapper[4744]: I1003 17:44:03.453699 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/crc-debug-wq9vc" Oct 03 17:44:03 crc kubenswrapper[4744]: I1003 17:44:03.495317 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzvc8\" (UniqueName: \"kubernetes.io/projected/c963dd0b-c455-4f0c-8aa5-c3f08cc3d644-kube-api-access-jzvc8\") on node \"crc\" DevicePath \"\"" Oct 03 17:44:03 crc kubenswrapper[4744]: I1003 17:44:03.495349 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c963dd0b-c455-4f0c-8aa5-c3f08cc3d644-host\") on node \"crc\" DevicePath \"\"" Oct 03 17:44:09 crc kubenswrapper[4744]: I1003 17:44:09.364648 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mhn28/crc-debug-wq9vc"] Oct 03 17:44:09 crc kubenswrapper[4744]: I1003 17:44:09.372061 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mhn28/crc-debug-wq9vc"] Oct 03 17:44:10 crc kubenswrapper[4744]: I1003 17:44:10.586356 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mhn28/crc-debug-dcpr7"] Oct 03 17:44:10 crc kubenswrapper[4744]: E1003 17:44:10.586966 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c963dd0b-c455-4f0c-8aa5-c3f08cc3d644" containerName="container-00" Oct 03 17:44:10 crc kubenswrapper[4744]: I1003 17:44:10.586978 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c963dd0b-c455-4f0c-8aa5-c3f08cc3d644" containerName="container-00" Oct 03 17:44:10 crc kubenswrapper[4744]: I1003 17:44:10.587185 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c963dd0b-c455-4f0c-8aa5-c3f08cc3d644" containerName="container-00" Oct 03 17:44:10 crc kubenswrapper[4744]: I1003 17:44:10.587751 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/crc-debug-dcpr7" Oct 03 17:44:10 crc kubenswrapper[4744]: I1003 17:44:10.714717 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4prkm\" (UniqueName: \"kubernetes.io/projected/d69494a0-6d93-459f-813c-c0d5f4d3e1b0-kube-api-access-4prkm\") pod \"crc-debug-dcpr7\" (UID: \"d69494a0-6d93-459f-813c-c0d5f4d3e1b0\") " pod="openshift-must-gather-mhn28/crc-debug-dcpr7" Oct 03 17:44:10 crc kubenswrapper[4744]: I1003 17:44:10.715219 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d69494a0-6d93-459f-813c-c0d5f4d3e1b0-host\") pod \"crc-debug-dcpr7\" (UID: \"d69494a0-6d93-459f-813c-c0d5f4d3e1b0\") " pod="openshift-must-gather-mhn28/crc-debug-dcpr7" Oct 03 17:44:10 crc kubenswrapper[4744]: I1003 17:44:10.818969 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4prkm\" (UniqueName: \"kubernetes.io/projected/d69494a0-6d93-459f-813c-c0d5f4d3e1b0-kube-api-access-4prkm\") pod \"crc-debug-dcpr7\" (UID: \"d69494a0-6d93-459f-813c-c0d5f4d3e1b0\") " pod="openshift-must-gather-mhn28/crc-debug-dcpr7" Oct 03 17:44:10 crc kubenswrapper[4744]: I1003 17:44:10.819114 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d69494a0-6d93-459f-813c-c0d5f4d3e1b0-host\") pod \"crc-debug-dcpr7\" (UID: \"d69494a0-6d93-459f-813c-c0d5f4d3e1b0\") " pod="openshift-must-gather-mhn28/crc-debug-dcpr7" Oct 03 17:44:10 crc kubenswrapper[4744]: I1003 17:44:10.819265 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d69494a0-6d93-459f-813c-c0d5f4d3e1b0-host\") pod \"crc-debug-dcpr7\" (UID: \"d69494a0-6d93-459f-813c-c0d5f4d3e1b0\") " pod="openshift-must-gather-mhn28/crc-debug-dcpr7" Oct 03 17:44:10 crc kubenswrapper[4744]: I1003 17:44:10.851837 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4prkm\" (UniqueName: \"kubernetes.io/projected/d69494a0-6d93-459f-813c-c0d5f4d3e1b0-kube-api-access-4prkm\") pod \"crc-debug-dcpr7\" (UID: \"d69494a0-6d93-459f-813c-c0d5f4d3e1b0\") " pod="openshift-must-gather-mhn28/crc-debug-dcpr7" Oct 03 17:44:10 crc kubenswrapper[4744]: I1003 17:44:10.906514 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/crc-debug-dcpr7" Oct 03 17:44:10 crc kubenswrapper[4744]: I1003 17:44:10.907591 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c963dd0b-c455-4f0c-8aa5-c3f08cc3d644" path="/var/lib/kubelet/pods/c963dd0b-c455-4f0c-8aa5-c3f08cc3d644/volumes" Oct 03 17:44:11 crc kubenswrapper[4744]: I1003 17:44:11.561673 4744 generic.go:334] "Generic (PLEG): container finished" podID="d69494a0-6d93-459f-813c-c0d5f4d3e1b0" containerID="16eed35584b2bc9a8dc6fdb18062d0a3fae9b688cabfa3a2efffc59648320fac" exitCode=0 Oct 03 17:44:11 crc kubenswrapper[4744]: I1003 17:44:11.561801 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mhn28/crc-debug-dcpr7" event={"ID":"d69494a0-6d93-459f-813c-c0d5f4d3e1b0","Type":"ContainerDied","Data":"16eed35584b2bc9a8dc6fdb18062d0a3fae9b688cabfa3a2efffc59648320fac"} Oct 03 17:44:11 crc kubenswrapper[4744]: I1003 17:44:11.561981 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mhn28/crc-debug-dcpr7" event={"ID":"d69494a0-6d93-459f-813c-c0d5f4d3e1b0","Type":"ContainerStarted","Data":"a69e73bf505d61d7deb19a829e8e25135ddab6d125b8b78e0628f494a11334ee"} Oct 03 17:44:11 crc kubenswrapper[4744]: I1003 17:44:11.619946 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mhn28/crc-debug-dcpr7"] Oct 03 17:44:11 crc kubenswrapper[4744]: I1003 17:44:11.635439 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mhn28/crc-debug-dcpr7"] Oct 03 17:44:11 crc kubenswrapper[4744]: I1003 17:44:11.893325 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:44:12 crc kubenswrapper[4744]: I1003 17:44:12.578408 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"c1441960a66e4ddfd471f97e8bfbe584c3f6a01e4d0501c4d81cf0e95cf9e35e"} Oct 03 17:44:12 crc kubenswrapper[4744]: I1003 17:44:12.726751 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/crc-debug-dcpr7" Oct 03 17:44:12 crc kubenswrapper[4744]: I1003 17:44:12.870701 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4prkm\" (UniqueName: \"kubernetes.io/projected/d69494a0-6d93-459f-813c-c0d5f4d3e1b0-kube-api-access-4prkm\") pod \"d69494a0-6d93-459f-813c-c0d5f4d3e1b0\" (UID: \"d69494a0-6d93-459f-813c-c0d5f4d3e1b0\") " Oct 03 17:44:12 crc kubenswrapper[4744]: I1003 17:44:12.870870 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d69494a0-6d93-459f-813c-c0d5f4d3e1b0-host\") pod \"d69494a0-6d93-459f-813c-c0d5f4d3e1b0\" (UID: \"d69494a0-6d93-459f-813c-c0d5f4d3e1b0\") " Oct 03 17:44:12 crc kubenswrapper[4744]: I1003 17:44:12.871624 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d69494a0-6d93-459f-813c-c0d5f4d3e1b0-host" (OuterVolumeSpecName: "host") pod "d69494a0-6d93-459f-813c-c0d5f4d3e1b0" (UID: "d69494a0-6d93-459f-813c-c0d5f4d3e1b0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 17:44:12 crc kubenswrapper[4744]: I1003 17:44:12.889176 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d69494a0-6d93-459f-813c-c0d5f4d3e1b0-kube-api-access-4prkm" (OuterVolumeSpecName: "kube-api-access-4prkm") pod "d69494a0-6d93-459f-813c-c0d5f4d3e1b0" (UID: "d69494a0-6d93-459f-813c-c0d5f4d3e1b0"). InnerVolumeSpecName "kube-api-access-4prkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:44:12 crc kubenswrapper[4744]: I1003 17:44:12.907003 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d69494a0-6d93-459f-813c-c0d5f4d3e1b0" path="/var/lib/kubelet/pods/d69494a0-6d93-459f-813c-c0d5f4d3e1b0/volumes" Oct 03 17:44:12 crc kubenswrapper[4744]: I1003 17:44:12.973552 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4prkm\" (UniqueName: \"kubernetes.io/projected/d69494a0-6d93-459f-813c-c0d5f4d3e1b0-kube-api-access-4prkm\") on node \"crc\" DevicePath \"\"" Oct 03 17:44:12 crc kubenswrapper[4744]: I1003 17:44:12.973579 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d69494a0-6d93-459f-813c-c0d5f4d3e1b0-host\") on node \"crc\" DevicePath \"\"" Oct 03 17:44:13 crc kubenswrapper[4744]: I1003 17:44:13.320618 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx_aaa2ef65-fb80-4f95-81fc-29841e8b6847/util/0.log" Oct 03 17:44:13 crc kubenswrapper[4744]: I1003 17:44:13.486392 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx_aaa2ef65-fb80-4f95-81fc-29841e8b6847/pull/0.log" Oct 03 17:44:13 crc kubenswrapper[4744]: I1003 17:44:13.486730 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx_aaa2ef65-fb80-4f95-81fc-29841e8b6847/pull/0.log" Oct 03 17:44:13 crc kubenswrapper[4744]: I1003 17:44:13.534706 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx_aaa2ef65-fb80-4f95-81fc-29841e8b6847/util/0.log" Oct 03 17:44:13 crc kubenswrapper[4744]: I1003 17:44:13.587238 4744 scope.go:117] "RemoveContainer" containerID="16eed35584b2bc9a8dc6fdb18062d0a3fae9b688cabfa3a2efffc59648320fac" Oct 03 17:44:13 crc kubenswrapper[4744]: I1003 17:44:13.587313 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/crc-debug-dcpr7" Oct 03 17:44:13 crc kubenswrapper[4744]: I1003 17:44:13.665277 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx_aaa2ef65-fb80-4f95-81fc-29841e8b6847/pull/0.log" Oct 03 17:44:13 crc kubenswrapper[4744]: I1003 17:44:13.670567 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx_aaa2ef65-fb80-4f95-81fc-29841e8b6847/extract/0.log" Oct 03 17:44:13 crc kubenswrapper[4744]: I1003 17:44:13.670871 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx_aaa2ef65-fb80-4f95-81fc-29841e8b6847/util/0.log" Oct 03 17:44:13 crc kubenswrapper[4744]: I1003 17:44:13.825646 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6c675fb79f-4st8l_d68c3557-6655-4449-8167-c28d1ccf442e/kube-rbac-proxy/0.log" Oct 03 17:44:13 crc kubenswrapper[4744]: I1003 17:44:13.845953 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79d68d6c85-4r792_829ba77b-d3cc-416c-8813-eeff7b6b60bd/kube-rbac-proxy/0.log" Oct 03 17:44:13 crc kubenswrapper[4744]: I1003 17:44:13.897555 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6c675fb79f-4st8l_d68c3557-6655-4449-8167-c28d1ccf442e/manager/0.log" Oct 03 17:44:14 crc kubenswrapper[4744]: I1003 17:44:14.056467 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79d68d6c85-4r792_829ba77b-d3cc-416c-8813-eeff7b6b60bd/manager/0.log" Oct 03 17:44:14 crc kubenswrapper[4744]: I1003 17:44:14.057894 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-bs7f6_19371562-3c8d-4dce-864c-0a37cbce2bf3/kube-rbac-proxy/0.log" Oct 03 17:44:14 crc kubenswrapper[4744]: I1003 17:44:14.087507 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-bs7f6_19371562-3c8d-4dce-864c-0a37cbce2bf3/manager/0.log" Oct 03 17:44:14 crc kubenswrapper[4744]: I1003 17:44:14.214708 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-846dff85b5-7qtkn_a38be30f-e956-4859-9610-a4d0735fe9d4/kube-rbac-proxy/0.log" Oct 03 17:44:14 crc kubenswrapper[4744]: I1003 17:44:14.309663 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-846dff85b5-7qtkn_a38be30f-e956-4859-9610-a4d0735fe9d4/manager/0.log" Oct 03 17:44:14 crc kubenswrapper[4744]: I1003 17:44:14.383477 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-599898f689-fxkn9_da0c0550-089f-4bd9-aaf8-960fcc75f019/kube-rbac-proxy/0.log" Oct 03 17:44:14 crc kubenswrapper[4744]: I1003 17:44:14.445035 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-599898f689-fxkn9_da0c0550-089f-4bd9-aaf8-960fcc75f019/manager/0.log" Oct 03 17:44:14 crc kubenswrapper[4744]: I1003 17:44:14.509788 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6769b867d9-n26jb_fd8ecb24-30b8-4012-846a-4618f178996b/kube-rbac-proxy/0.log" Oct 03 17:44:14 crc kubenswrapper[4744]: I1003 17:44:14.600149 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6769b867d9-n26jb_fd8ecb24-30b8-4012-846a-4618f178996b/manager/0.log" Oct 03 17:44:14 crc kubenswrapper[4744]: I1003 17:44:14.726696 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5fbf469cd7-4jdmt_3bc36b86-4888-4341-b52d-b9b73013f9ad/kube-rbac-proxy/0.log" Oct 03 17:44:14 crc kubenswrapper[4744]: I1003 17:44:14.840858 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5fbf469cd7-4jdmt_3bc36b86-4888-4341-b52d-b9b73013f9ad/manager/0.log" Oct 03 17:44:14 crc kubenswrapper[4744]: I1003 17:44:14.900688 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-84bc9db6cc-tb5k2_23519c20-6aec-4461-bea9-adaf88e87ab3/kube-rbac-proxy/0.log" Oct 03 17:44:14 crc kubenswrapper[4744]: I1003 17:44:14.975127 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-84bc9db6cc-tb5k2_23519c20-6aec-4461-bea9-adaf88e87ab3/manager/0.log" Oct 03 17:44:15 crc kubenswrapper[4744]: I1003 17:44:15.055580 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7f55849f88-r8tdt_a862283f-4243-42bd-bb2c-5b2392d657ac/kube-rbac-proxy/0.log" Oct 03 17:44:15 crc kubenswrapper[4744]: I1003 17:44:15.158469 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7f55849f88-r8tdt_a862283f-4243-42bd-bb2c-5b2392d657ac/manager/0.log" Oct 03 17:44:15 crc kubenswrapper[4744]: I1003 17:44:15.288234 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6fd6854b49-zd5fc_be9443a6-46c8-4374-8caa-8040ae117981/kube-rbac-proxy/0.log" Oct 03 17:44:15 crc kubenswrapper[4744]: I1003 17:44:15.305670 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6fd6854b49-zd5fc_be9443a6-46c8-4374-8caa-8040ae117981/manager/0.log" Oct 03 17:44:15 crc kubenswrapper[4744]: I1003 17:44:15.422395 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5c468bf4d4-kzsg2_d34995fe-607b-46ae-a2eb-c91d9c244415/kube-rbac-proxy/0.log" Oct 03 17:44:15 crc kubenswrapper[4744]: I1003 17:44:15.498942 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5c468bf4d4-kzsg2_d34995fe-607b-46ae-a2eb-c91d9c244415/manager/0.log" Oct 03 17:44:15 crc kubenswrapper[4744]: I1003 17:44:15.551821 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6574bf987d-xxdpr_2832b841-b383-4eed-a8ce-71062269238a/kube-rbac-proxy/0.log" Oct 03 17:44:15 crc kubenswrapper[4744]: I1003 17:44:15.632634 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6574bf987d-xxdpr_2832b841-b383-4eed-a8ce-71062269238a/manager/0.log" Oct 03 17:44:15 crc kubenswrapper[4744]: I1003 17:44:15.776910 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-555c7456bd-6xg8v_438b2904-4b85-4d82-b883-f5788d4caab9/kube-rbac-proxy/0.log" Oct 03 17:44:15 crc kubenswrapper[4744]: I1003 17:44:15.803157 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-555c7456bd-6xg8v_438b2904-4b85-4d82-b883-f5788d4caab9/manager/0.log" Oct 03 17:44:15 crc kubenswrapper[4744]: I1003 17:44:15.916271 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-59d6cfdf45-xr6db_d34377f2-1b67-4c62-9602-ddb93f0cb4d7/kube-rbac-proxy/0.log" Oct 03 17:44:15 crc kubenswrapper[4744]: I1003 17:44:15.971051 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-59d6cfdf45-xr6db_d34377f2-1b67-4c62-9602-ddb93f0cb4d7/manager/0.log" Oct 03 17:44:16 crc kubenswrapper[4744]: I1003 17:44:16.020338 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6f64c4d678v879q_911d5ffd-d447-4b8a-a177-7fc738d2aa35/kube-rbac-proxy/0.log" Oct 03 17:44:16 crc kubenswrapper[4744]: I1003 17:44:16.098091 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6f64c4d678v879q_911d5ffd-d447-4b8a-a177-7fc738d2aa35/manager/0.log" Oct 03 17:44:16 crc kubenswrapper[4744]: I1003 17:44:16.185554 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7758d4b55d-6xczl_4717fade-30fb-4f78-b7fb-dec2799d9b8c/kube-rbac-proxy/0.log" Oct 03 17:44:16 crc kubenswrapper[4744]: I1003 17:44:16.383860 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-d94d76459-krksv_df419359-3ee4-4c00-bb37-c7387b19492a/kube-rbac-proxy/0.log" Oct 03 17:44:16 crc kubenswrapper[4744]: I1003 17:44:16.536612 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-d94d76459-krksv_df419359-3ee4-4c00-bb37-c7387b19492a/operator/0.log" Oct 03 17:44:16 crc kubenswrapper[4744]: I1003 17:44:16.613225 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-n2bfv_d8b895bc-8e19-4b92-999f-d579c1d51e51/registry-server/0.log" Oct 03 17:44:16 crc kubenswrapper[4744]: I1003 17:44:16.750078 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-688db7b6c7-jhqdv_8be63e93-5f9d-4ffe-9ab2-d5b794c34f9b/kube-rbac-proxy/0.log" Oct 03 17:44:16 crc kubenswrapper[4744]: I1003 17:44:16.804962 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-688db7b6c7-jhqdv_8be63e93-5f9d-4ffe-9ab2-d5b794c34f9b/manager/0.log" Oct 03 17:44:16 crc kubenswrapper[4744]: I1003 17:44:16.849563 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-7d8bb7f44c-jktv9_39de7155-617f-444a-bb16-0cb7fb229eb7/kube-rbac-proxy/0.log" Oct 03 17:44:17 crc kubenswrapper[4744]: I1003 17:44:17.033724 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw_aef71e84-6035-42fb-a16f-c0e97876f1c0/operator/0.log" Oct 03 17:44:17 crc kubenswrapper[4744]: I1003 17:44:17.066522 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-7d8bb7f44c-jktv9_39de7155-617f-444a-bb16-0cb7fb229eb7/manager/0.log" Oct 03 17:44:17 crc kubenswrapper[4744]: I1003 17:44:17.215789 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-psqc4_a6943fb4-ab7a-445f-b4d1-be9089562b77/kube-rbac-proxy/0.log" Oct 03 17:44:17 crc kubenswrapper[4744]: I1003 17:44:17.271237 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-psqc4_a6943fb4-ab7a-445f-b4d1-be9089562b77/manager/0.log" Oct 03 17:44:17 crc kubenswrapper[4744]: I1003 17:44:17.338618 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5db5cf686f-7qgsm_09c1f75f-d699-4675-b78d-a37353acd758/kube-rbac-proxy/0.log" Oct 03 17:44:17 crc kubenswrapper[4744]: I1003 17:44:17.355821 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7758d4b55d-6xczl_4717fade-30fb-4f78-b7fb-dec2799d9b8c/manager/0.log" Oct 03 17:44:17 crc kubenswrapper[4744]: I1003 17:44:17.506239 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5db5cf686f-7qgsm_09c1f75f-d699-4675-b78d-a37353acd758/manager/0.log" Oct 03 17:44:17 crc kubenswrapper[4744]: I1003 17:44:17.523343 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-ksw89_1d753e81-b04a-4216-991f-256a2ab6ac94/kube-rbac-proxy/0.log" Oct 03 17:44:17 crc kubenswrapper[4744]: I1003 17:44:17.527634 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-ksw89_1d753e81-b04a-4216-991f-256a2ab6ac94/manager/0.log" Oct 03 17:44:17 crc kubenswrapper[4744]: I1003 17:44:17.653417 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-fcd7d9895-wjl4s_1aa4d11e-aebd-477b-98e1-a05ea74a7677/kube-rbac-proxy/0.log" Oct 03 17:44:18 crc kubenswrapper[4744]: I1003 17:44:18.630128 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-fcd7d9895-wjl4s_1aa4d11e-aebd-477b-98e1-a05ea74a7677/manager/0.log" Oct 03 17:44:35 crc kubenswrapper[4744]: I1003 17:44:35.395965 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-qmzm9_c41f7473-52d9-4d32-9364-ed61efb65af4/control-plane-machine-set-operator/0.log" Oct 03 17:44:35 crc kubenswrapper[4744]: I1003 17:44:35.592818 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-btrfh_23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c/machine-api-operator/0.log" Oct 03 17:44:35 crc kubenswrapper[4744]: I1003 17:44:35.606621 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-btrfh_23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c/kube-rbac-proxy/0.log" Oct 03 17:44:48 crc kubenswrapper[4744]: I1003 17:44:48.208187 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-m8bhx_15fdf250-5959-4502-b7be-4d7af4eff05d/cert-manager-controller/0.log" Oct 03 17:44:48 crc kubenswrapper[4744]: I1003 17:44:48.921858 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-xjmdh_f62bcdb5-0b26-4be1-af80-c413f76ce50c/cert-manager-webhook/0.log" Oct 03 17:44:48 crc kubenswrapper[4744]: I1003 17:44:48.923618 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-xp59p_862c46f4-2e50-4556-8405-bfcdb337ed95/cert-manager-cainjector/0.log" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.144049 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj"] Oct 03 17:45:00 crc kubenswrapper[4744]: E1003 17:45:00.145039 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d69494a0-6d93-459f-813c-c0d5f4d3e1b0" containerName="container-00" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.145056 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d69494a0-6d93-459f-813c-c0d5f4d3e1b0" containerName="container-00" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.145340 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d69494a0-6d93-459f-813c-c0d5f4d3e1b0" containerName="container-00" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.146198 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.148467 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.148483 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.160895 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj"] Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.272371 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bad798b8-5b08-431d-bee3-c7897d034228-secret-volume\") pod \"collect-profiles-29325225-dd9jj\" (UID: \"bad798b8-5b08-431d-bee3-c7897d034228\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.272418 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bad798b8-5b08-431d-bee3-c7897d034228-config-volume\") pod \"collect-profiles-29325225-dd9jj\" (UID: \"bad798b8-5b08-431d-bee3-c7897d034228\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.272592 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pshw2\" (UniqueName: \"kubernetes.io/projected/bad798b8-5b08-431d-bee3-c7897d034228-kube-api-access-pshw2\") pod \"collect-profiles-29325225-dd9jj\" (UID: \"bad798b8-5b08-431d-bee3-c7897d034228\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.374874 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bad798b8-5b08-431d-bee3-c7897d034228-secret-volume\") pod \"collect-profiles-29325225-dd9jj\" (UID: \"bad798b8-5b08-431d-bee3-c7897d034228\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.374951 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bad798b8-5b08-431d-bee3-c7897d034228-config-volume\") pod \"collect-profiles-29325225-dd9jj\" (UID: \"bad798b8-5b08-431d-bee3-c7897d034228\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.375061 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pshw2\" (UniqueName: \"kubernetes.io/projected/bad798b8-5b08-431d-bee3-c7897d034228-kube-api-access-pshw2\") pod \"collect-profiles-29325225-dd9jj\" (UID: \"bad798b8-5b08-431d-bee3-c7897d034228\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.376079 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bad798b8-5b08-431d-bee3-c7897d034228-config-volume\") pod \"collect-profiles-29325225-dd9jj\" (UID: \"bad798b8-5b08-431d-bee3-c7897d034228\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.716636 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bad798b8-5b08-431d-bee3-c7897d034228-secret-volume\") pod \"collect-profiles-29325225-dd9jj\" (UID: \"bad798b8-5b08-431d-bee3-c7897d034228\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.727117 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pshw2\" (UniqueName: \"kubernetes.io/projected/bad798b8-5b08-431d-bee3-c7897d034228-kube-api-access-pshw2\") pod \"collect-profiles-29325225-dd9jj\" (UID: \"bad798b8-5b08-431d-bee3-c7897d034228\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" Oct 03 17:45:00 crc kubenswrapper[4744]: I1003 17:45:00.764384 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" Oct 03 17:45:01 crc kubenswrapper[4744]: I1003 17:45:01.250409 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj"] Oct 03 17:45:01 crc kubenswrapper[4744]: E1003 17:45:01.886524 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbad798b8_5b08_431d_bee3_c7897d034228.slice/crio-conmon-4991a699b75f034777b66f56924e3d523df523cc8329fdd9aefee1f05bab6c50.scope\": RecentStats: unable to find data in memory cache]" Oct 03 17:45:02 crc kubenswrapper[4744]: I1003 17:45:02.059448 4744 generic.go:334] "Generic (PLEG): container finished" podID="bad798b8-5b08-431d-bee3-c7897d034228" containerID="4991a699b75f034777b66f56924e3d523df523cc8329fdd9aefee1f05bab6c50" exitCode=0 Oct 03 17:45:02 crc kubenswrapper[4744]: I1003 17:45:02.059504 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" event={"ID":"bad798b8-5b08-431d-bee3-c7897d034228","Type":"ContainerDied","Data":"4991a699b75f034777b66f56924e3d523df523cc8329fdd9aefee1f05bab6c50"} Oct 03 17:45:02 crc kubenswrapper[4744]: I1003 17:45:02.059532 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" event={"ID":"bad798b8-5b08-431d-bee3-c7897d034228","Type":"ContainerStarted","Data":"8e73cb881e4c80b689397a63413fb1258f884c2edf0d69c34c5f43e93079fc39"} Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.017844 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-cvn5h_d0fe8f50-20f3-483f-8904-8f084fba1fb1/nmstate-handler/0.log" Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.062224 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-587v7_c5f8637e-bf33-4533-970d-a21b8837ce78/nmstate-console-plugin/0.log" Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.210003 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-pdpss_4039e497-2af2-483a-8a21-14f14df6df33/kube-rbac-proxy/0.log" Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.233285 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-pdpss_4039e497-2af2-483a-8a21-14f14df6df33/nmstate-metrics/0.log" Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.361932 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-qd582_af44daee-90fc-4ccd-a4dd-ae2513f0fe2c/nmstate-operator/0.log" Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.441286 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.443337 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-dmz4f_63038935-53b5-4371-ab75-80e6c1522980/nmstate-webhook/0.log" Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.540889 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bad798b8-5b08-431d-bee3-c7897d034228-config-volume\") pod \"bad798b8-5b08-431d-bee3-c7897d034228\" (UID: \"bad798b8-5b08-431d-bee3-c7897d034228\") " Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.541126 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pshw2\" (UniqueName: \"kubernetes.io/projected/bad798b8-5b08-431d-bee3-c7897d034228-kube-api-access-pshw2\") pod \"bad798b8-5b08-431d-bee3-c7897d034228\" (UID: \"bad798b8-5b08-431d-bee3-c7897d034228\") " Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.541269 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bad798b8-5b08-431d-bee3-c7897d034228-secret-volume\") pod \"bad798b8-5b08-431d-bee3-c7897d034228\" (UID: \"bad798b8-5b08-431d-bee3-c7897d034228\") " Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.541597 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bad798b8-5b08-431d-bee3-c7897d034228-config-volume" (OuterVolumeSpecName: "config-volume") pod "bad798b8-5b08-431d-bee3-c7897d034228" (UID: "bad798b8-5b08-431d-bee3-c7897d034228"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.541709 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bad798b8-5b08-431d-bee3-c7897d034228-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.547294 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad798b8-5b08-431d-bee3-c7897d034228-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bad798b8-5b08-431d-bee3-c7897d034228" (UID: "bad798b8-5b08-431d-bee3-c7897d034228"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.547340 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bad798b8-5b08-431d-bee3-c7897d034228-kube-api-access-pshw2" (OuterVolumeSpecName: "kube-api-access-pshw2") pod "bad798b8-5b08-431d-bee3-c7897d034228" (UID: "bad798b8-5b08-431d-bee3-c7897d034228"). InnerVolumeSpecName "kube-api-access-pshw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.643969 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bad798b8-5b08-431d-bee3-c7897d034228-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 17:45:03 crc kubenswrapper[4744]: I1003 17:45:03.644005 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pshw2\" (UniqueName: \"kubernetes.io/projected/bad798b8-5b08-431d-bee3-c7897d034228-kube-api-access-pshw2\") on node \"crc\" DevicePath \"\"" Oct 03 17:45:04 crc kubenswrapper[4744]: I1003 17:45:04.077955 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" event={"ID":"bad798b8-5b08-431d-bee3-c7897d034228","Type":"ContainerDied","Data":"8e73cb881e4c80b689397a63413fb1258f884c2edf0d69c34c5f43e93079fc39"} Oct 03 17:45:04 crc kubenswrapper[4744]: I1003 17:45:04.078014 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e73cb881e4c80b689397a63413fb1258f884c2edf0d69c34c5f43e93079fc39" Oct 03 17:45:04 crc kubenswrapper[4744]: I1003 17:45:04.078089 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325225-dd9jj" Oct 03 17:45:04 crc kubenswrapper[4744]: I1003 17:45:04.516100 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj"] Oct 03 17:45:04 crc kubenswrapper[4744]: I1003 17:45:04.526357 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325180-8znfj"] Oct 03 17:45:04 crc kubenswrapper[4744]: I1003 17:45:04.904906 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8701dc0d-5341-4ed0-a151-840095189550" path="/var/lib/kubelet/pods/8701dc0d-5341-4ed0-a151-840095189550/volumes" Oct 03 17:45:17 crc kubenswrapper[4744]: I1003 17:45:17.225534 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-vhk6s_5f04d600-8bb7-4420-aa5f-c1b51a0a4e16/kube-rbac-proxy/0.log" Oct 03 17:45:17 crc kubenswrapper[4744]: I1003 17:45:17.284495 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-vhk6s_5f04d600-8bb7-4420-aa5f-c1b51a0a4e16/controller/0.log" Oct 03 17:45:17 crc kubenswrapper[4744]: I1003 17:45:17.382885 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-frr-files/0.log" Oct 03 17:45:17 crc kubenswrapper[4744]: I1003 17:45:17.600379 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-frr-files/0.log" Oct 03 17:45:17 crc kubenswrapper[4744]: I1003 17:45:17.619818 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-reloader/0.log" Oct 03 17:45:17 crc kubenswrapper[4744]: I1003 17:45:17.638901 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-reloader/0.log" Oct 03 17:45:17 crc kubenswrapper[4744]: I1003 17:45:17.659055 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-metrics/0.log" Oct 03 17:45:17 crc kubenswrapper[4744]: I1003 17:45:17.800645 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-frr-files/0.log" Oct 03 17:45:17 crc kubenswrapper[4744]: I1003 17:45:17.828437 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-metrics/0.log" Oct 03 17:45:17 crc kubenswrapper[4744]: I1003 17:45:17.840945 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-metrics/0.log" Oct 03 17:45:17 crc kubenswrapper[4744]: I1003 17:45:17.840980 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-reloader/0.log" Oct 03 17:45:18 crc kubenswrapper[4744]: I1003 17:45:18.061763 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-reloader/0.log" Oct 03 17:45:18 crc kubenswrapper[4744]: I1003 17:45:18.071303 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-metrics/0.log" Oct 03 17:45:18 crc kubenswrapper[4744]: I1003 17:45:18.072040 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-frr-files/0.log" Oct 03 17:45:18 crc kubenswrapper[4744]: I1003 17:45:18.102025 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/controller/0.log" Oct 03 17:45:18 crc kubenswrapper[4744]: I1003 17:45:18.227754 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/kube-rbac-proxy/0.log" Oct 03 17:45:18 crc kubenswrapper[4744]: I1003 17:45:18.230470 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/frr-metrics/0.log" Oct 03 17:45:18 crc kubenswrapper[4744]: I1003 17:45:18.368914 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/kube-rbac-proxy-frr/0.log" Oct 03 17:45:18 crc kubenswrapper[4744]: I1003 17:45:18.405303 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/reloader/0.log" Oct 03 17:45:18 crc kubenswrapper[4744]: I1003 17:45:18.574565 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-xsl5f_703cac15-6674-4529-9a6b-75f646a84d3e/frr-k8s-webhook-server/0.log" Oct 03 17:45:18 crc kubenswrapper[4744]: I1003 17:45:18.669993 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6cbff679b9-cdsl8_8776a23e-e0c0-4c8d-84e8-d34d5036e3c6/manager/0.log" Oct 03 17:45:18 crc kubenswrapper[4744]: I1003 17:45:18.869321 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-86c9f64ddf-r6kfg_259f45bb-e4ab-45f4-9d77-5e788361c35a/webhook-server/0.log" Oct 03 17:45:19 crc kubenswrapper[4744]: I1003 17:45:19.063836 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5qx66_0d474bc7-3b94-40d8-91d6-db976b2975d7/kube-rbac-proxy/0.log" Oct 03 17:45:19 crc kubenswrapper[4744]: I1003 17:45:19.579241 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5qx66_0d474bc7-3b94-40d8-91d6-db976b2975d7/speaker/0.log" Oct 03 17:45:19 crc kubenswrapper[4744]: I1003 17:45:19.779955 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/frr/0.log" Oct 03 17:45:32 crc kubenswrapper[4744]: I1003 17:45:32.750157 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh_c795770c-99b7-4077-a8b4-f50b678f32b8/util/0.log" Oct 03 17:45:32 crc kubenswrapper[4744]: I1003 17:45:32.950185 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh_c795770c-99b7-4077-a8b4-f50b678f32b8/pull/0.log" Oct 03 17:45:32 crc kubenswrapper[4744]: I1003 17:45:32.968737 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh_c795770c-99b7-4077-a8b4-f50b678f32b8/pull/0.log" Oct 03 17:45:32 crc kubenswrapper[4744]: I1003 17:45:32.970275 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh_c795770c-99b7-4077-a8b4-f50b678f32b8/util/0.log" Oct 03 17:45:33 crc kubenswrapper[4744]: I1003 17:45:33.138616 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh_c795770c-99b7-4077-a8b4-f50b678f32b8/pull/0.log" Oct 03 17:45:33 crc kubenswrapper[4744]: I1003 17:45:33.149615 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh_c795770c-99b7-4077-a8b4-f50b678f32b8/util/0.log" Oct 03 17:45:33 crc kubenswrapper[4744]: I1003 17:45:33.157924 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh_c795770c-99b7-4077-a8b4-f50b678f32b8/extract/0.log" Oct 03 17:45:33 crc kubenswrapper[4744]: I1003 17:45:33.304784 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcfbk_f6fb66df-a4db-417b-90fc-c779777883be/extract-utilities/0.log" Oct 03 17:45:33 crc kubenswrapper[4744]: I1003 17:45:33.443944 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcfbk_f6fb66df-a4db-417b-90fc-c779777883be/extract-content/0.log" Oct 03 17:45:33 crc kubenswrapper[4744]: I1003 17:45:33.473254 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcfbk_f6fb66df-a4db-417b-90fc-c779777883be/extract-utilities/0.log" Oct 03 17:45:33 crc kubenswrapper[4744]: I1003 17:45:33.499466 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcfbk_f6fb66df-a4db-417b-90fc-c779777883be/extract-content/0.log" Oct 03 17:45:33 crc kubenswrapper[4744]: I1003 17:45:33.641573 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcfbk_f6fb66df-a4db-417b-90fc-c779777883be/extract-utilities/0.log" Oct 03 17:45:33 crc kubenswrapper[4744]: I1003 17:45:33.690770 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcfbk_f6fb66df-a4db-417b-90fc-c779777883be/extract-content/0.log" Oct 03 17:45:34 crc kubenswrapper[4744]: I1003 17:45:34.175054 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcfbk_f6fb66df-a4db-417b-90fc-c779777883be/registry-server/0.log" Oct 03 17:45:34 crc kubenswrapper[4744]: I1003 17:45:34.360107 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7v9wq_6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2/extract-utilities/0.log" Oct 03 17:45:34 crc kubenswrapper[4744]: I1003 17:45:34.568513 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7v9wq_6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2/extract-content/0.log" Oct 03 17:45:34 crc kubenswrapper[4744]: I1003 17:45:34.571537 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7v9wq_6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2/extract-utilities/0.log" Oct 03 17:45:34 crc kubenswrapper[4744]: I1003 17:45:34.583342 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7v9wq_6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2/extract-content/0.log" Oct 03 17:45:34 crc kubenswrapper[4744]: I1003 17:45:34.733711 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7v9wq_6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2/extract-utilities/0.log" Oct 03 17:45:34 crc kubenswrapper[4744]: I1003 17:45:34.787043 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7v9wq_6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2/extract-content/0.log" Oct 03 17:45:34 crc kubenswrapper[4744]: I1003 17:45:34.993311 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd_6a7d3bdf-d461-4fcb-810c-421130e8d524/util/0.log" Oct 03 17:45:35 crc kubenswrapper[4744]: I1003 17:45:35.157095 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd_6a7d3bdf-d461-4fcb-810c-421130e8d524/pull/0.log" Oct 03 17:45:35 crc kubenswrapper[4744]: I1003 17:45:35.185852 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd_6a7d3bdf-d461-4fcb-810c-421130e8d524/util/0.log" Oct 03 17:45:35 crc kubenswrapper[4744]: I1003 17:45:35.252054 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd_6a7d3bdf-d461-4fcb-810c-421130e8d524/pull/0.log" Oct 03 17:45:35 crc kubenswrapper[4744]: I1003 17:45:35.437771 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd_6a7d3bdf-d461-4fcb-810c-421130e8d524/util/0.log" Oct 03 17:45:35 crc kubenswrapper[4744]: I1003 17:45:35.468226 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7v9wq_6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2/registry-server/0.log" Oct 03 17:45:35 crc kubenswrapper[4744]: I1003 17:45:35.484357 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd_6a7d3bdf-d461-4fcb-810c-421130e8d524/pull/0.log" Oct 03 17:45:35 crc kubenswrapper[4744]: I1003 17:45:35.497128 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd_6a7d3bdf-d461-4fcb-810c-421130e8d524/extract/0.log" Oct 03 17:45:36 crc kubenswrapper[4744]: I1003 17:45:36.239090 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xr8s7_bddf7be1-c68d-42ab-853e-153710114dd4/marketplace-operator/0.log" Oct 03 17:45:36 crc kubenswrapper[4744]: I1003 17:45:36.273546 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wmqbq_86654f3a-97fa-456a-b651-93f8ec174725/extract-utilities/0.log" Oct 03 17:45:36 crc kubenswrapper[4744]: I1003 17:45:36.441092 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wmqbq_86654f3a-97fa-456a-b651-93f8ec174725/extract-utilities/0.log" Oct 03 17:45:36 crc kubenswrapper[4744]: I1003 17:45:36.444979 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wmqbq_86654f3a-97fa-456a-b651-93f8ec174725/extract-content/0.log" Oct 03 17:45:36 crc kubenswrapper[4744]: I1003 17:45:36.447699 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wmqbq_86654f3a-97fa-456a-b651-93f8ec174725/extract-content/0.log" Oct 03 17:45:36 crc kubenswrapper[4744]: I1003 17:45:36.643818 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wmqbq_86654f3a-97fa-456a-b651-93f8ec174725/extract-content/0.log" Oct 03 17:45:36 crc kubenswrapper[4744]: I1003 17:45:36.667557 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wmqbq_86654f3a-97fa-456a-b651-93f8ec174725/extract-utilities/0.log" Oct 03 17:45:36 crc kubenswrapper[4744]: I1003 17:45:36.734317 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l8gwc_83f32f6a-6664-4e8f-8676-cefcc31b6e6c/extract-utilities/0.log" Oct 03 17:45:36 crc kubenswrapper[4744]: I1003 17:45:36.832248 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wmqbq_86654f3a-97fa-456a-b651-93f8ec174725/registry-server/0.log" Oct 03 17:45:36 crc kubenswrapper[4744]: I1003 17:45:36.903327 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l8gwc_83f32f6a-6664-4e8f-8676-cefcc31b6e6c/extract-content/0.log" Oct 03 17:45:36 crc kubenswrapper[4744]: I1003 17:45:36.932780 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l8gwc_83f32f6a-6664-4e8f-8676-cefcc31b6e6c/extract-content/0.log" Oct 03 17:45:36 crc kubenswrapper[4744]: I1003 17:45:36.945345 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l8gwc_83f32f6a-6664-4e8f-8676-cefcc31b6e6c/extract-utilities/0.log" Oct 03 17:45:37 crc kubenswrapper[4744]: I1003 17:45:37.084056 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l8gwc_83f32f6a-6664-4e8f-8676-cefcc31b6e6c/extract-utilities/0.log" Oct 03 17:45:37 crc kubenswrapper[4744]: I1003 17:45:37.128825 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l8gwc_83f32f6a-6664-4e8f-8676-cefcc31b6e6c/extract-content/0.log" Oct 03 17:45:37 crc kubenswrapper[4744]: I1003 17:45:37.684966 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l8gwc_83f32f6a-6664-4e8f-8676-cefcc31b6e6c/registry-server/0.log" Oct 03 17:45:41 crc kubenswrapper[4744]: I1003 17:45:41.231769 4744 scope.go:117] "RemoveContainer" containerID="8f6a6cd09f3369bd5c121f96147e9edd40ef64ec9fa852ab872b20dee3985ecd" Oct 03 17:46:11 crc kubenswrapper[4744]: E1003 17:46:11.318986 4744 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.132:45474->38.102.83.132:38347: read tcp 38.102.83.132:45474->38.102.83.132:38347: read: connection reset by peer Oct 03 17:46:34 crc kubenswrapper[4744]: I1003 17:46:34.668689 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:46:34 crc kubenswrapper[4744]: I1003 17:46:34.669192 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:47:04 crc kubenswrapper[4744]: I1003 17:47:04.668340 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:47:04 crc kubenswrapper[4744]: I1003 17:47:04.669087 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:47:34 crc kubenswrapper[4744]: I1003 17:47:34.668148 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:47:34 crc kubenswrapper[4744]: I1003 17:47:34.668663 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:47:34 crc kubenswrapper[4744]: I1003 17:47:34.668700 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 17:47:34 crc kubenswrapper[4744]: I1003 17:47:34.669441 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c1441960a66e4ddfd471f97e8bfbe584c3f6a01e4d0501c4d81cf0e95cf9e35e"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 17:47:34 crc kubenswrapper[4744]: I1003 17:47:34.669503 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://c1441960a66e4ddfd471f97e8bfbe584c3f6a01e4d0501c4d81cf0e95cf9e35e" gracePeriod=600 Oct 03 17:47:35 crc kubenswrapper[4744]: I1003 17:47:35.676920 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="c1441960a66e4ddfd471f97e8bfbe584c3f6a01e4d0501c4d81cf0e95cf9e35e" exitCode=0 Oct 03 17:47:35 crc kubenswrapper[4744]: I1003 17:47:35.676996 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"c1441960a66e4ddfd471f97e8bfbe584c3f6a01e4d0501c4d81cf0e95cf9e35e"} Oct 03 17:47:35 crc kubenswrapper[4744]: I1003 17:47:35.677323 4744 scope.go:117] "RemoveContainer" containerID="8c56d82b7c800c2bd7279a4d16e12d53f8b5495033b7e6f76b92067dae507537" Oct 03 17:47:36 crc kubenswrapper[4744]: I1003 17:47:36.703279 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2"} Oct 03 17:48:02 crc kubenswrapper[4744]: I1003 17:48:02.994531 4744 generic.go:334] "Generic (PLEG): container finished" podID="c7c5232f-a750-4320-95b6-df7b9866babe" containerID="3bd725223847135cc3e7488666794ed27f33c99e5549007dee5277afe20ca140" exitCode=0 Oct 03 17:48:02 crc kubenswrapper[4744]: I1003 17:48:02.994654 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mhn28/must-gather-pzkn4" event={"ID":"c7c5232f-a750-4320-95b6-df7b9866babe","Type":"ContainerDied","Data":"3bd725223847135cc3e7488666794ed27f33c99e5549007dee5277afe20ca140"} Oct 03 17:48:02 crc kubenswrapper[4744]: I1003 17:48:02.996066 4744 scope.go:117] "RemoveContainer" containerID="3bd725223847135cc3e7488666794ed27f33c99e5549007dee5277afe20ca140" Oct 03 17:48:03 crc kubenswrapper[4744]: I1003 17:48:03.595124 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mhn28_must-gather-pzkn4_c7c5232f-a750-4320-95b6-df7b9866babe/gather/0.log" Oct 03 17:48:12 crc kubenswrapper[4744]: I1003 17:48:12.551893 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mhn28/must-gather-pzkn4"] Oct 03 17:48:12 crc kubenswrapper[4744]: I1003 17:48:12.552616 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mhn28/must-gather-pzkn4" podUID="c7c5232f-a750-4320-95b6-df7b9866babe" containerName="copy" containerID="cri-o://a2a8f55a9d23bbf8785a3ba76c418a62f7df52b1f839687bc6660f09b9a8edce" gracePeriod=2 Oct 03 17:48:12 crc kubenswrapper[4744]: I1003 17:48:12.569763 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mhn28/must-gather-pzkn4"] Oct 03 17:48:12 crc kubenswrapper[4744]: I1003 17:48:12.997362 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mhn28_must-gather-pzkn4_c7c5232f-a750-4320-95b6-df7b9866babe/copy/0.log" Oct 03 17:48:12 crc kubenswrapper[4744]: I1003 17:48:12.998205 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/must-gather-pzkn4" Oct 03 17:48:13 crc kubenswrapper[4744]: I1003 17:48:13.070646 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmmdk\" (UniqueName: \"kubernetes.io/projected/c7c5232f-a750-4320-95b6-df7b9866babe-kube-api-access-dmmdk\") pod \"c7c5232f-a750-4320-95b6-df7b9866babe\" (UID: \"c7c5232f-a750-4320-95b6-df7b9866babe\") " Oct 03 17:48:13 crc kubenswrapper[4744]: I1003 17:48:13.070727 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c7c5232f-a750-4320-95b6-df7b9866babe-must-gather-output\") pod \"c7c5232f-a750-4320-95b6-df7b9866babe\" (UID: \"c7c5232f-a750-4320-95b6-df7b9866babe\") " Oct 03 17:48:13 crc kubenswrapper[4744]: I1003 17:48:13.076680 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7c5232f-a750-4320-95b6-df7b9866babe-kube-api-access-dmmdk" (OuterVolumeSpecName: "kube-api-access-dmmdk") pod "c7c5232f-a750-4320-95b6-df7b9866babe" (UID: "c7c5232f-a750-4320-95b6-df7b9866babe"). InnerVolumeSpecName "kube-api-access-dmmdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:48:13 crc kubenswrapper[4744]: I1003 17:48:13.103840 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mhn28_must-gather-pzkn4_c7c5232f-a750-4320-95b6-df7b9866babe/copy/0.log" Oct 03 17:48:13 crc kubenswrapper[4744]: I1003 17:48:13.104416 4744 generic.go:334] "Generic (PLEG): container finished" podID="c7c5232f-a750-4320-95b6-df7b9866babe" containerID="a2a8f55a9d23bbf8785a3ba76c418a62f7df52b1f839687bc6660f09b9a8edce" exitCode=143 Oct 03 17:48:13 crc kubenswrapper[4744]: I1003 17:48:13.104488 4744 scope.go:117] "RemoveContainer" containerID="a2a8f55a9d23bbf8785a3ba76c418a62f7df52b1f839687bc6660f09b9a8edce" Oct 03 17:48:13 crc kubenswrapper[4744]: I1003 17:48:13.104527 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mhn28/must-gather-pzkn4" Oct 03 17:48:13 crc kubenswrapper[4744]: I1003 17:48:13.158171 4744 scope.go:117] "RemoveContainer" containerID="3bd725223847135cc3e7488666794ed27f33c99e5549007dee5277afe20ca140" Oct 03 17:48:13 crc kubenswrapper[4744]: I1003 17:48:13.174817 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmmdk\" (UniqueName: \"kubernetes.io/projected/c7c5232f-a750-4320-95b6-df7b9866babe-kube-api-access-dmmdk\") on node \"crc\" DevicePath \"\"" Oct 03 17:48:13 crc kubenswrapper[4744]: I1003 17:48:13.249083 4744 scope.go:117] "RemoveContainer" containerID="a2a8f55a9d23bbf8785a3ba76c418a62f7df52b1f839687bc6660f09b9a8edce" Oct 03 17:48:13 crc kubenswrapper[4744]: E1003 17:48:13.249650 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2a8f55a9d23bbf8785a3ba76c418a62f7df52b1f839687bc6660f09b9a8edce\": container with ID starting with a2a8f55a9d23bbf8785a3ba76c418a62f7df52b1f839687bc6660f09b9a8edce not found: ID does not exist" containerID="a2a8f55a9d23bbf8785a3ba76c418a62f7df52b1f839687bc6660f09b9a8edce" Oct 03 17:48:13 crc kubenswrapper[4744]: I1003 17:48:13.249695 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2a8f55a9d23bbf8785a3ba76c418a62f7df52b1f839687bc6660f09b9a8edce"} err="failed to get container status \"a2a8f55a9d23bbf8785a3ba76c418a62f7df52b1f839687bc6660f09b9a8edce\": rpc error: code = NotFound desc = could not find container \"a2a8f55a9d23bbf8785a3ba76c418a62f7df52b1f839687bc6660f09b9a8edce\": container with ID starting with a2a8f55a9d23bbf8785a3ba76c418a62f7df52b1f839687bc6660f09b9a8edce not found: ID does not exist" Oct 03 17:48:13 crc kubenswrapper[4744]: I1003 17:48:13.249720 4744 scope.go:117] "RemoveContainer" containerID="3bd725223847135cc3e7488666794ed27f33c99e5549007dee5277afe20ca140" Oct 03 17:48:13 crc kubenswrapper[4744]: E1003 17:48:13.250267 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bd725223847135cc3e7488666794ed27f33c99e5549007dee5277afe20ca140\": container with ID starting with 3bd725223847135cc3e7488666794ed27f33c99e5549007dee5277afe20ca140 not found: ID does not exist" containerID="3bd725223847135cc3e7488666794ed27f33c99e5549007dee5277afe20ca140" Oct 03 17:48:13 crc kubenswrapper[4744]: I1003 17:48:13.250321 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bd725223847135cc3e7488666794ed27f33c99e5549007dee5277afe20ca140"} err="failed to get container status \"3bd725223847135cc3e7488666794ed27f33c99e5549007dee5277afe20ca140\": rpc error: code = NotFound desc = could not find container \"3bd725223847135cc3e7488666794ed27f33c99e5549007dee5277afe20ca140\": container with ID starting with 3bd725223847135cc3e7488666794ed27f33c99e5549007dee5277afe20ca140 not found: ID does not exist" Oct 03 17:48:13 crc kubenswrapper[4744]: I1003 17:48:13.273667 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7c5232f-a750-4320-95b6-df7b9866babe-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "c7c5232f-a750-4320-95b6-df7b9866babe" (UID: "c7c5232f-a750-4320-95b6-df7b9866babe"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:48:13 crc kubenswrapper[4744]: I1003 17:48:13.276517 4744 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c7c5232f-a750-4320-95b6-df7b9866babe-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 03 17:48:14 crc kubenswrapper[4744]: I1003 17:48:14.908881 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7c5232f-a750-4320-95b6-df7b9866babe" path="/var/lib/kubelet/pods/c7c5232f-a750-4320-95b6-df7b9866babe/volumes" Oct 03 17:48:36 crc kubenswrapper[4744]: I1003 17:48:36.741348 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mgmdg/must-gather-5h7jg"] Oct 03 17:48:36 crc kubenswrapper[4744]: E1003 17:48:36.742634 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7c5232f-a750-4320-95b6-df7b9866babe" containerName="copy" Oct 03 17:48:36 crc kubenswrapper[4744]: I1003 17:48:36.742653 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7c5232f-a750-4320-95b6-df7b9866babe" containerName="copy" Oct 03 17:48:36 crc kubenswrapper[4744]: E1003 17:48:36.742678 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad798b8-5b08-431d-bee3-c7897d034228" containerName="collect-profiles" Oct 03 17:48:36 crc kubenswrapper[4744]: I1003 17:48:36.742687 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad798b8-5b08-431d-bee3-c7897d034228" containerName="collect-profiles" Oct 03 17:48:36 crc kubenswrapper[4744]: E1003 17:48:36.742735 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7c5232f-a750-4320-95b6-df7b9866babe" containerName="gather" Oct 03 17:48:36 crc kubenswrapper[4744]: I1003 17:48:36.742744 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7c5232f-a750-4320-95b6-df7b9866babe" containerName="gather" Oct 03 17:48:36 crc kubenswrapper[4744]: I1003 17:48:36.742989 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad798b8-5b08-431d-bee3-c7897d034228" containerName="collect-profiles" Oct 03 17:48:36 crc kubenswrapper[4744]: I1003 17:48:36.743015 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7c5232f-a750-4320-95b6-df7b9866babe" containerName="copy" Oct 03 17:48:36 crc kubenswrapper[4744]: I1003 17:48:36.743038 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7c5232f-a750-4320-95b6-df7b9866babe" containerName="gather" Oct 03 17:48:36 crc kubenswrapper[4744]: I1003 17:48:36.744394 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/must-gather-5h7jg" Oct 03 17:48:36 crc kubenswrapper[4744]: I1003 17:48:36.755846 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mgmdg"/"openshift-service-ca.crt" Oct 03 17:48:36 crc kubenswrapper[4744]: I1003 17:48:36.756076 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mgmdg"/"kube-root-ca.crt" Oct 03 17:48:36 crc kubenswrapper[4744]: I1003 17:48:36.756252 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mgmdg"/"default-dockercfg-mnfgc" Oct 03 17:48:36 crc kubenswrapper[4744]: I1003 17:48:36.757863 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mgmdg/must-gather-5h7jg"] Oct 03 17:48:36 crc kubenswrapper[4744]: I1003 17:48:36.916259 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3e32a7a9-b8ca-4900-a343-22c70c44f69c-must-gather-output\") pod \"must-gather-5h7jg\" (UID: \"3e32a7a9-b8ca-4900-a343-22c70c44f69c\") " pod="openshift-must-gather-mgmdg/must-gather-5h7jg" Oct 03 17:48:36 crc kubenswrapper[4744]: I1003 17:48:36.916451 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sklhb\" (UniqueName: \"kubernetes.io/projected/3e32a7a9-b8ca-4900-a343-22c70c44f69c-kube-api-access-sklhb\") pod \"must-gather-5h7jg\" (UID: \"3e32a7a9-b8ca-4900-a343-22c70c44f69c\") " pod="openshift-must-gather-mgmdg/must-gather-5h7jg" Oct 03 17:48:37 crc kubenswrapper[4744]: I1003 17:48:37.019156 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sklhb\" (UniqueName: \"kubernetes.io/projected/3e32a7a9-b8ca-4900-a343-22c70c44f69c-kube-api-access-sklhb\") pod \"must-gather-5h7jg\" (UID: \"3e32a7a9-b8ca-4900-a343-22c70c44f69c\") " pod="openshift-must-gather-mgmdg/must-gather-5h7jg" Oct 03 17:48:37 crc kubenswrapper[4744]: I1003 17:48:37.019432 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3e32a7a9-b8ca-4900-a343-22c70c44f69c-must-gather-output\") pod \"must-gather-5h7jg\" (UID: \"3e32a7a9-b8ca-4900-a343-22c70c44f69c\") " pod="openshift-must-gather-mgmdg/must-gather-5h7jg" Oct 03 17:48:37 crc kubenswrapper[4744]: I1003 17:48:37.019826 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3e32a7a9-b8ca-4900-a343-22c70c44f69c-must-gather-output\") pod \"must-gather-5h7jg\" (UID: \"3e32a7a9-b8ca-4900-a343-22c70c44f69c\") " pod="openshift-must-gather-mgmdg/must-gather-5h7jg" Oct 03 17:48:37 crc kubenswrapper[4744]: I1003 17:48:37.039787 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sklhb\" (UniqueName: \"kubernetes.io/projected/3e32a7a9-b8ca-4900-a343-22c70c44f69c-kube-api-access-sklhb\") pod \"must-gather-5h7jg\" (UID: \"3e32a7a9-b8ca-4900-a343-22c70c44f69c\") " pod="openshift-must-gather-mgmdg/must-gather-5h7jg" Oct 03 17:48:37 crc kubenswrapper[4744]: I1003 17:48:37.073833 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/must-gather-5h7jg" Oct 03 17:48:37 crc kubenswrapper[4744]: I1003 17:48:37.376907 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mgmdg/must-gather-5h7jg"] Oct 03 17:48:38 crc kubenswrapper[4744]: I1003 17:48:38.386740 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgmdg/must-gather-5h7jg" event={"ID":"3e32a7a9-b8ca-4900-a343-22c70c44f69c","Type":"ContainerStarted","Data":"ab51547dffcfedac9a57e4c70d3ccc11d31f6b850f608e1635db38e434486693"} Oct 03 17:48:38 crc kubenswrapper[4744]: I1003 17:48:38.387063 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgmdg/must-gather-5h7jg" event={"ID":"3e32a7a9-b8ca-4900-a343-22c70c44f69c","Type":"ContainerStarted","Data":"3e7292228fd4f55ef95277d281f282b4ec45ebc06d8ddfca00fbf108c4be6cdd"} Oct 03 17:48:38 crc kubenswrapper[4744]: I1003 17:48:38.387075 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgmdg/must-gather-5h7jg" event={"ID":"3e32a7a9-b8ca-4900-a343-22c70c44f69c","Type":"ContainerStarted","Data":"90a396b341e1a7a7c2e33aaec9f464a9ac3a244535e62e2e303e683d3dd1b9c1"} Oct 03 17:48:38 crc kubenswrapper[4744]: I1003 17:48:38.400033 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mgmdg/must-gather-5h7jg" podStartSLOduration=2.400015717 podStartE2EDuration="2.400015717s" podCreationTimestamp="2025-10-03 17:48:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 17:48:38.399686669 +0000 UTC m=+5044.679562565" watchObservedRunningTime="2025-10-03 17:48:38.400015717 +0000 UTC m=+5044.679891613" Oct 03 17:48:41 crc kubenswrapper[4744]: I1003 17:48:41.476471 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mgmdg/crc-debug-62tst"] Oct 03 17:48:41 crc kubenswrapper[4744]: I1003 17:48:41.479831 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/crc-debug-62tst" Oct 03 17:48:41 crc kubenswrapper[4744]: I1003 17:48:41.605714 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8-host\") pod \"crc-debug-62tst\" (UID: \"7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8\") " pod="openshift-must-gather-mgmdg/crc-debug-62tst" Oct 03 17:48:41 crc kubenswrapper[4744]: I1003 17:48:41.606170 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8fsv\" (UniqueName: \"kubernetes.io/projected/7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8-kube-api-access-q8fsv\") pod \"crc-debug-62tst\" (UID: \"7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8\") " pod="openshift-must-gather-mgmdg/crc-debug-62tst" Oct 03 17:48:41 crc kubenswrapper[4744]: I1003 17:48:41.707509 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8fsv\" (UniqueName: \"kubernetes.io/projected/7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8-kube-api-access-q8fsv\") pod \"crc-debug-62tst\" (UID: \"7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8\") " pod="openshift-must-gather-mgmdg/crc-debug-62tst" Oct 03 17:48:41 crc kubenswrapper[4744]: I1003 17:48:41.707675 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8-host\") pod \"crc-debug-62tst\" (UID: \"7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8\") " pod="openshift-must-gather-mgmdg/crc-debug-62tst" Oct 03 17:48:41 crc kubenswrapper[4744]: I1003 17:48:41.707794 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8-host\") pod \"crc-debug-62tst\" (UID: \"7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8\") " pod="openshift-must-gather-mgmdg/crc-debug-62tst" Oct 03 17:48:41 crc kubenswrapper[4744]: I1003 17:48:41.725438 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8fsv\" (UniqueName: \"kubernetes.io/projected/7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8-kube-api-access-q8fsv\") pod \"crc-debug-62tst\" (UID: \"7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8\") " pod="openshift-must-gather-mgmdg/crc-debug-62tst" Oct 03 17:48:41 crc kubenswrapper[4744]: I1003 17:48:41.801578 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/crc-debug-62tst" Oct 03 17:48:42 crc kubenswrapper[4744]: I1003 17:48:42.441570 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgmdg/crc-debug-62tst" event={"ID":"7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8","Type":"ContainerStarted","Data":"c96e625d8863c983d205e97d9d8eea65ab82994189ee690be72f45096916ac2a"} Oct 03 17:48:43 crc kubenswrapper[4744]: I1003 17:48:43.451611 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgmdg/crc-debug-62tst" event={"ID":"7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8","Type":"ContainerStarted","Data":"7f5e513fdbb52c3368824c4108beb7191433497048df2c33f4b5428f50aa3408"} Oct 03 17:48:43 crc kubenswrapper[4744]: I1003 17:48:43.469587 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mgmdg/crc-debug-62tst" podStartSLOduration=2.469564569 podStartE2EDuration="2.469564569s" podCreationTimestamp="2025-10-03 17:48:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 17:48:43.464635077 +0000 UTC m=+5049.744510993" watchObservedRunningTime="2025-10-03 17:48:43.469564569 +0000 UTC m=+5049.749440485" Oct 03 17:49:49 crc kubenswrapper[4744]: I1003 17:49:49.962250 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7bf8c46bb6-7m9v6_068cb298-fa9c-4e22-92cc-0aa31310f185/barbican-api-log/0.log" Oct 03 17:49:49 crc kubenswrapper[4744]: I1003 17:49:49.998281 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7bf8c46bb6-7m9v6_068cb298-fa9c-4e22-92cc-0aa31310f185/barbican-api/0.log" Oct 03 17:49:50 crc kubenswrapper[4744]: I1003 17:49:50.207663 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6dfd6c4c58-rr7fr_6b4d5e00-5174-4dc3-b68e-185dedb764f1/barbican-keystone-listener/0.log" Oct 03 17:49:50 crc kubenswrapper[4744]: I1003 17:49:50.910759 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c46748695-jn5bb_9c69aee3-5a63-4011-b4ff-0afca3c16880/barbican-worker/0.log" Oct 03 17:49:51 crc kubenswrapper[4744]: I1003 17:49:51.006742 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6dfd6c4c58-rr7fr_6b4d5e00-5174-4dc3-b68e-185dedb764f1/barbican-keystone-listener-log/0.log" Oct 03 17:49:51 crc kubenswrapper[4744]: I1003 17:49:51.090152 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c46748695-jn5bb_9c69aee3-5a63-4011-b4ff-0afca3c16880/barbican-worker-log/0.log" Oct 03 17:49:51 crc kubenswrapper[4744]: I1003 17:49:51.245217 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-87gx9_b860d7da-e83d-4a61-aa1d-ba22cf817d8d/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:49:51 crc kubenswrapper[4744]: I1003 17:49:51.387065 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_13a29aaf-7761-4250-b9fd-bb69661f450e/ceilometer-central-agent/0.log" Oct 03 17:49:51 crc kubenswrapper[4744]: I1003 17:49:51.460328 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_13a29aaf-7761-4250-b9fd-bb69661f450e/ceilometer-notification-agent/0.log" Oct 03 17:49:51 crc kubenswrapper[4744]: I1003 17:49:51.508893 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_13a29aaf-7761-4250-b9fd-bb69661f450e/proxy-httpd/0.log" Oct 03 17:49:51 crc kubenswrapper[4744]: I1003 17:49:51.573178 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_13a29aaf-7761-4250-b9fd-bb69661f450e/sg-core/0.log" Oct 03 17:49:51 crc kubenswrapper[4744]: I1003 17:49:51.831640 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph_af690e2a-fac7-4b2b-9ca0-4473e9f6989d/ceph/0.log" Oct 03 17:49:52 crc kubenswrapper[4744]: I1003 17:49:52.084630 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_cee08598-c007-4049-8b75-d4be9c1b0796/cinder-api/0.log" Oct 03 17:49:52 crc kubenswrapper[4744]: I1003 17:49:52.195799 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_cee08598-c007-4049-8b75-d4be9c1b0796/cinder-api-log/0.log" Oct 03 17:49:52 crc kubenswrapper[4744]: I1003 17:49:52.875793 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_33874615-0e1f-4743-aa84-d262b6eab68c/probe/0.log" Oct 03 17:49:53 crc kubenswrapper[4744]: I1003 17:49:53.207858 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_008783bb-cc87-416a-9823-787ff756a3c9/cinder-scheduler/0.log" Oct 03 17:49:53 crc kubenswrapper[4744]: I1003 17:49:53.208080 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_008783bb-cc87-416a-9823-787ff756a3c9/probe/0.log" Oct 03 17:49:53 crc kubenswrapper[4744]: I1003 17:49:53.534059 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_146d5ce6-9a87-450e-b56c-371f2aa7af06/probe/0.log" Oct 03 17:49:53 crc kubenswrapper[4744]: I1003 17:49:53.750603 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-2bqlg_4fe42e89-74c8-4ea1-b1f5-af28a032f3ce/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:49:53 crc kubenswrapper[4744]: I1003 17:49:53.979913 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-h4w9c_2ac77289-8098-4cdd-892f-a56916c886e6/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:49:54 crc kubenswrapper[4744]: I1003 17:49:54.143826 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-hx4hz_af14e991-2a54-4228-aa49-a8ce99d8eab2/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:49:54 crc kubenswrapper[4744]: I1003 17:49:54.381538 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5bb847fbb7-w6gpl_0bbdd34c-0174-4307-a779-0f8520f6e19a/init/0.log" Oct 03 17:49:54 crc kubenswrapper[4744]: I1003 17:49:54.623455 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5bb847fbb7-w6gpl_0bbdd34c-0174-4307-a779-0f8520f6e19a/init/0.log" Oct 03 17:49:54 crc kubenswrapper[4744]: I1003 17:49:54.804200 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5bb847fbb7-w6gpl_0bbdd34c-0174-4307-a779-0f8520f6e19a/dnsmasq-dns/0.log" Oct 03 17:49:55 crc kubenswrapper[4744]: I1003 17:49:55.019131 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-bs6g8_b2145bd1-f9de-41e1-b1c5-39f0dcd0c8a1/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:49:55 crc kubenswrapper[4744]: I1003 17:49:55.227886 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_411dd208-5058-4f51-b5ab-4e4f8c77cd8e/glance-httpd/0.log" Oct 03 17:49:55 crc kubenswrapper[4744]: I1003 17:49:55.257671 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_411dd208-5058-4f51-b5ab-4e4f8c77cd8e/glance-log/0.log" Oct 03 17:49:55 crc kubenswrapper[4744]: I1003 17:49:55.539600 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b7679ad4-70b7-4058-a2e6-c4dba8a409e9/glance-httpd/0.log" Oct 03 17:49:55 crc kubenswrapper[4744]: I1003 17:49:55.654271 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_b7679ad4-70b7-4058-a2e6-c4dba8a409e9/glance-log/0.log" Oct 03 17:49:55 crc kubenswrapper[4744]: I1003 17:49:55.970511 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-85d8c4f7b-9pn28_39c711a0-3cea-4091-94a0-ec9df6e6a2d8/horizon/0.log" Oct 03 17:49:56 crc kubenswrapper[4744]: I1003 17:49:56.461858 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-z4jhc_3aa5f64b-0451-4034-822c-d462f01e7429/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:49:56 crc kubenswrapper[4744]: I1003 17:49:56.891760 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_146d5ce6-9a87-450e-b56c-371f2aa7af06/cinder-volume/0.log" Oct 03 17:49:56 crc kubenswrapper[4744]: I1003 17:49:56.906286 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-85d8c4f7b-9pn28_39c711a0-3cea-4091-94a0-ec9df6e6a2d8/horizon-log/0.log" Oct 03 17:49:56 crc kubenswrapper[4744]: I1003 17:49:56.943633 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-jbg68_42a1ea96-7b56-4594-811d-7043e8640e39/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:49:57 crc kubenswrapper[4744]: I1003 17:49:57.361204 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29325181-pc2dm_513c18a1-42b9-43f1-bb0c-4a99a3c45ea6/keystone-cron/0.log" Oct 03 17:49:57 crc kubenswrapper[4744]: I1003 17:49:57.520890 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_6c69a90c-bccf-4209-94e1-fe80d4914377/kube-state-metrics/0.log" Oct 03 17:49:57 crc kubenswrapper[4744]: I1003 17:49:57.805163 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-tjpz5_78b5d479-1346-45e4-96f9-310b6b32ec50/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:49:58 crc kubenswrapper[4744]: I1003 17:49:58.343909 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_33874615-0e1f-4743-aa84-d262b6eab68c/cinder-backup/0.log" Oct 03 17:49:58 crc kubenswrapper[4744]: I1003 17:49:58.828536 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_f392aa4a-226f-4009-acbb-d66b4aaeb95e/manila-api/0.log" Oct 03 17:49:59 crc kubenswrapper[4744]: I1003 17:49:59.066917 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_75efca1b-e080-4620-ad1b-8e1bcc3ba3ba/probe/0.log" Oct 03 17:49:59 crc kubenswrapper[4744]: I1003 17:49:59.155237 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_75efca1b-e080-4620-ad1b-8e1bcc3ba3ba/manila-scheduler/0.log" Oct 03 17:49:59 crc kubenswrapper[4744]: I1003 17:49:59.344138 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_f392aa4a-226f-4009-acbb-d66b4aaeb95e/manila-api-log/0.log" Oct 03 17:49:59 crc kubenswrapper[4744]: I1003 17:49:59.511606 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_3c066c82-d9c3-40bc-805d-16a45e328e49/probe/0.log" Oct 03 17:49:59 crc kubenswrapper[4744]: I1003 17:49:59.803535 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_3c066c82-d9c3-40bc-805d-16a45e328e49/manila-share/0.log" Oct 03 17:50:01 crc kubenswrapper[4744]: I1003 17:50:01.000975 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6bc59588f-s74gr_ec5a5a4c-4896-4629-be5d-2b96f2729eaf/neutron-httpd/0.log" Oct 03 17:50:01 crc kubenswrapper[4744]: I1003 17:50:01.477078 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-66dvz_a1fb9143-c0e9-44b2-b80d-ff1ed760a115/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:50:01 crc kubenswrapper[4744]: I1003 17:50:01.589730 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6589b6d4db-rbbtp_6e8f7c20-0249-41e8-9cc6-915cf0762b3a/keystone-api/0.log" Oct 03 17:50:01 crc kubenswrapper[4744]: I1003 17:50:01.606104 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6bc59588f-s74gr_ec5a5a4c-4896-4629-be5d-2b96f2729eaf/neutron-api/0.log" Oct 03 17:50:03 crc kubenswrapper[4744]: I1003 17:50:03.596144 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_da322563-8083-4e70-b339-a0c68c23e0c3/nova-cell0-conductor-conductor/0.log" Oct 03 17:50:04 crc kubenswrapper[4744]: I1003 17:50:04.317532 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_74a40f5d-b401-437e-9e4c-38e1d3459e40/nova-cell1-conductor-conductor/0.log" Oct 03 17:50:04 crc kubenswrapper[4744]: I1003 17:50:04.560644 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92/nova-api-log/0.log" Oct 03 17:50:04 crc kubenswrapper[4744]: I1003 17:50:04.667909 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:50:04 crc kubenswrapper[4744]: I1003 17:50:04.667968 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:50:04 crc kubenswrapper[4744]: I1003 17:50:04.925274 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_8005560a-4d21-4e19-9e47-17a703457765/nova-cell1-novncproxy-novncproxy/0.log" Oct 03 17:50:05 crc kubenswrapper[4744]: I1003 17:50:05.116691 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-z279b_710d1522-5c50-4e85-bb48-a099cfbb892d/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:50:05 crc kubenswrapper[4744]: I1003 17:50:05.386983 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_76b30dc4-9f12-483e-874e-6654e45e521a/nova-metadata-log/0.log" Oct 03 17:50:05 crc kubenswrapper[4744]: I1003 17:50:05.472245 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_5bcdc1c5-737a-48bd-b5a2-e5d5ea352e92/nova-api-api/0.log" Oct 03 17:50:06 crc kubenswrapper[4744]: I1003 17:50:06.093792 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8/mysql-bootstrap/0.log" Oct 03 17:50:06 crc kubenswrapper[4744]: I1003 17:50:06.149221 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1ae57486-d534-4c67-a235-218fe8293726/nova-scheduler-scheduler/0.log" Oct 03 17:50:06 crc kubenswrapper[4744]: I1003 17:50:06.296612 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8/mysql-bootstrap/0.log" Oct 03 17:50:06 crc kubenswrapper[4744]: I1003 17:50:06.383614 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_71ec1585-ca83-4fd6-9bb2-f6e92b29cbe8/galera/0.log" Oct 03 17:50:06 crc kubenswrapper[4744]: I1003 17:50:06.614695 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7c8031bb-6ec3-42b3-ace7-80e25071e230/mysql-bootstrap/0.log" Oct 03 17:50:06 crc kubenswrapper[4744]: I1003 17:50:06.815873 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7c8031bb-6ec3-42b3-ace7-80e25071e230/mysql-bootstrap/0.log" Oct 03 17:50:06 crc kubenswrapper[4744]: I1003 17:50:06.845545 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_7c8031bb-6ec3-42b3-ace7-80e25071e230/galera/0.log" Oct 03 17:50:07 crc kubenswrapper[4744]: I1003 17:50:07.026569 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_38e0f66b-95e2-4ff4-a09a-1106a0eaa4f1/openstackclient/0.log" Oct 03 17:50:07 crc kubenswrapper[4744]: I1003 17:50:07.226775 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4jn8x_bc8fb837-d1e5-4953-ae6c-fbee5f867e49/ovn-controller/0.log" Oct 03 17:50:07 crc kubenswrapper[4744]: I1003 17:50:07.485482 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_76b30dc4-9f12-483e-874e-6654e45e521a/nova-metadata-metadata/0.log" Oct 03 17:50:07 crc kubenswrapper[4744]: I1003 17:50:07.582481 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-mgnmg_02bac763-b977-4d19-a6fd-300b0b2cefc3/openstack-network-exporter/0.log" Oct 03 17:50:07 crc kubenswrapper[4744]: I1003 17:50:07.679372 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-99k4w_77d29fac-2896-4f1a-a193-f4b559f1e9b6/ovsdb-server-init/0.log" Oct 03 17:50:07 crc kubenswrapper[4744]: I1003 17:50:07.926307 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-99k4w_77d29fac-2896-4f1a-a193-f4b559f1e9b6/ovsdb-server-init/0.log" Oct 03 17:50:08 crc kubenswrapper[4744]: I1003 17:50:08.017150 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-99k4w_77d29fac-2896-4f1a-a193-f4b559f1e9b6/ovsdb-server/0.log" Oct 03 17:50:08 crc kubenswrapper[4744]: I1003 17:50:08.020661 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-99k4w_77d29fac-2896-4f1a-a193-f4b559f1e9b6/ovs-vswitchd/0.log" Oct 03 17:50:08 crc kubenswrapper[4744]: I1003 17:50:08.305471 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-8x7d2_6c864721-c102-4044-bd81-74f3fbe29040/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:50:08 crc kubenswrapper[4744]: I1003 17:50:08.426772 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_09af7ba6-34b5-4454-9dae-5bcc301ed20c/openstack-network-exporter/0.log" Oct 03 17:50:08 crc kubenswrapper[4744]: I1003 17:50:08.517522 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_09af7ba6-34b5-4454-9dae-5bcc301ed20c/ovn-northd/0.log" Oct 03 17:50:08 crc kubenswrapper[4744]: I1003 17:50:08.643661 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_803d0f2b-9d7d-4d98-909c-18ffdec34aed/openstack-network-exporter/0.log" Oct 03 17:50:08 crc kubenswrapper[4744]: I1003 17:50:08.687417 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_803d0f2b-9d7d-4d98-909c-18ffdec34aed/ovsdbserver-nb/0.log" Oct 03 17:50:08 crc kubenswrapper[4744]: I1003 17:50:08.920478 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7c9e2064-beda-46d7-9221-06b2bbb375eb/openstack-network-exporter/0.log" Oct 03 17:50:08 crc kubenswrapper[4744]: I1003 17:50:08.946040 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7c9e2064-beda-46d7-9221-06b2bbb375eb/ovsdbserver-sb/0.log" Oct 03 17:50:09 crc kubenswrapper[4744]: I1003 17:50:09.540888 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-774fb64f68-qgzkk_0d2b9c37-6d79-420d-bc4e-3d5888b5d61e/placement-api/0.log" Oct 03 17:50:09 crc kubenswrapper[4744]: I1003 17:50:09.603312 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d34a40f-b279-4d50-88c9-bb7776fc678b/setup-container/0.log" Oct 03 17:50:09 crc kubenswrapper[4744]: I1003 17:50:09.657965 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-774fb64f68-qgzkk_0d2b9c37-6d79-420d-bc4e-3d5888b5d61e/placement-log/0.log" Oct 03 17:50:10 crc kubenswrapper[4744]: I1003 17:50:10.280445 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d34a40f-b279-4d50-88c9-bb7776fc678b/setup-container/0.log" Oct 03 17:50:10 crc kubenswrapper[4744]: I1003 17:50:10.406186 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d34a40f-b279-4d50-88c9-bb7776fc678b/rabbitmq/0.log" Oct 03 17:50:10 crc kubenswrapper[4744]: I1003 17:50:10.539554 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_52159513-8a05-44c2-89f2-fb6d08e50b36/setup-container/0.log" Oct 03 17:50:10 crc kubenswrapper[4744]: I1003 17:50:10.642371 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_52159513-8a05-44c2-89f2-fb6d08e50b36/setup-container/0.log" Oct 03 17:50:10 crc kubenswrapper[4744]: I1003 17:50:10.764009 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_52159513-8a05-44c2-89f2-fb6d08e50b36/rabbitmq/0.log" Oct 03 17:50:10 crc kubenswrapper[4744]: I1003 17:50:10.875728 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-jcxpc_0f862434-d64e-4c02-b8de-a120185d0053/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:50:11 crc kubenswrapper[4744]: I1003 17:50:11.035669 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-tr7p9_9f6a23c2-45cf-4443-b1ef-befcea950ae8/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:50:11 crc kubenswrapper[4744]: I1003 17:50:11.226596 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-7n5s7_962e48ea-a515-44c9-b2c2-2e9705a31455/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:50:11 crc kubenswrapper[4744]: I1003 17:50:11.702711 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-mf5dj_1b94f190-a56d-420d-9558-7ec1116b90b9/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:50:11 crc kubenswrapper[4744]: I1003 17:50:11.934288 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8nbg9_1700fe09-1277-4cda-89dc-03d220941b9a/ssh-known-hosts-edpm-deployment/0.log" Oct 03 17:50:12 crc kubenswrapper[4744]: I1003 17:50:12.204147 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-689c67db59-x766r_1c64dd03-9cb9-402f-9697-a23538c9ebc5/proxy-httpd/0.log" Oct 03 17:50:12 crc kubenswrapper[4744]: I1003 17:50:12.220015 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-689c67db59-x766r_1c64dd03-9cb9-402f-9697-a23538c9ebc5/proxy-server/0.log" Oct 03 17:50:12 crc kubenswrapper[4744]: I1003 17:50:12.406009 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-dsc89_0b88692d-9c4d-4c67-b70f-921810ca71e1/swift-ring-rebalance/0.log" Oct 03 17:50:12 crc kubenswrapper[4744]: I1003 17:50:12.522206 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/account-auditor/0.log" Oct 03 17:50:12 crc kubenswrapper[4744]: I1003 17:50:12.639905 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/account-reaper/0.log" Oct 03 17:50:12 crc kubenswrapper[4744]: I1003 17:50:12.734553 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/account-replicator/0.log" Oct 03 17:50:12 crc kubenswrapper[4744]: I1003 17:50:12.766615 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/account-server/0.log" Oct 03 17:50:12 crc kubenswrapper[4744]: I1003 17:50:12.848387 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/container-auditor/0.log" Oct 03 17:50:12 crc kubenswrapper[4744]: I1003 17:50:12.953043 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/container-replicator/0.log" Oct 03 17:50:12 crc kubenswrapper[4744]: I1003 17:50:12.984764 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/container-server/0.log" Oct 03 17:50:13 crc kubenswrapper[4744]: I1003 17:50:13.060444 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/container-updater/0.log" Oct 03 17:50:13 crc kubenswrapper[4744]: I1003 17:50:13.189336 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/object-auditor/0.log" Oct 03 17:50:13 crc kubenswrapper[4744]: I1003 17:50:13.220295 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/object-expirer/0.log" Oct 03 17:50:13 crc kubenswrapper[4744]: I1003 17:50:13.327002 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/object-replicator/0.log" Oct 03 17:50:13 crc kubenswrapper[4744]: I1003 17:50:13.400595 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/object-server/0.log" Oct 03 17:50:13 crc kubenswrapper[4744]: I1003 17:50:13.435658 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/object-updater/0.log" Oct 03 17:50:13 crc kubenswrapper[4744]: I1003 17:50:13.564237 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/rsync/0.log" Oct 03 17:50:13 crc kubenswrapper[4744]: I1003 17:50:13.585528 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_cc1aa9df-ee4d-4cd6-b35a-914b9656c0b7/memcached/0.log" Oct 03 17:50:13 crc kubenswrapper[4744]: I1003 17:50:13.621719 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_133725f5-8ac3-4f45-b6c3-f7aea913711d/swift-recon-cron/0.log" Oct 03 17:50:13 crc kubenswrapper[4744]: I1003 17:50:13.766663 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-xnfzn_4d4a0574-77fd-40fe-aa97-c4a26255473e/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:50:13 crc kubenswrapper[4744]: I1003 17:50:13.877663 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_595f783e-db00-49b2-968f-79163001ffdd/tempest-tests-tempest-tests-runner/0.log" Oct 03 17:50:13 crc kubenswrapper[4744]: I1003 17:50:13.911716 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_812d60e3-44e3-4f22-9afa-448f0ac216ef/test-operator-logs-container/0.log" Oct 03 17:50:14 crc kubenswrapper[4744]: I1003 17:50:14.052738 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-nr4r7_d8f4bcbd-4d4e-4b36-89c6-aabcf12f8f5b/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 17:50:19 crc kubenswrapper[4744]: I1003 17:50:19.969331 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wdjqp"] Oct 03 17:50:19 crc kubenswrapper[4744]: I1003 17:50:19.976641 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:20 crc kubenswrapper[4744]: I1003 17:50:20.030141 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wdjqp"] Oct 03 17:50:20 crc kubenswrapper[4744]: I1003 17:50:20.085117 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-catalog-content\") pod \"certified-operators-wdjqp\" (UID: \"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1\") " pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:20 crc kubenswrapper[4744]: I1003 17:50:20.085172 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-utilities\") pod \"certified-operators-wdjqp\" (UID: \"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1\") " pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:20 crc kubenswrapper[4744]: I1003 17:50:20.085231 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwd2q\" (UniqueName: \"kubernetes.io/projected/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-kube-api-access-gwd2q\") pod \"certified-operators-wdjqp\" (UID: \"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1\") " pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:20 crc kubenswrapper[4744]: I1003 17:50:20.187690 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-catalog-content\") pod \"certified-operators-wdjqp\" (UID: \"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1\") " pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:20 crc kubenswrapper[4744]: I1003 17:50:20.187760 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-utilities\") pod \"certified-operators-wdjqp\" (UID: \"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1\") " pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:20 crc kubenswrapper[4744]: I1003 17:50:20.187828 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwd2q\" (UniqueName: \"kubernetes.io/projected/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-kube-api-access-gwd2q\") pod \"certified-operators-wdjqp\" (UID: \"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1\") " pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:20 crc kubenswrapper[4744]: I1003 17:50:20.188658 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-catalog-content\") pod \"certified-operators-wdjqp\" (UID: \"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1\") " pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:20 crc kubenswrapper[4744]: I1003 17:50:20.188936 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-utilities\") pod \"certified-operators-wdjqp\" (UID: \"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1\") " pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:20 crc kubenswrapper[4744]: I1003 17:50:20.214178 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwd2q\" (UniqueName: \"kubernetes.io/projected/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-kube-api-access-gwd2q\") pod \"certified-operators-wdjqp\" (UID: \"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1\") " pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:20 crc kubenswrapper[4744]: I1003 17:50:20.315522 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:20 crc kubenswrapper[4744]: I1003 17:50:20.885469 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wdjqp"] Oct 03 17:50:21 crc kubenswrapper[4744]: I1003 17:50:21.360086 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdjqp" event={"ID":"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1","Type":"ContainerStarted","Data":"70f8181937c0cd6063d9e695a535adb56bdf256a4d1fd066d5ab1e96fc2d4417"} Oct 03 17:50:22 crc kubenswrapper[4744]: I1003 17:50:22.377348 4744 generic.go:334] "Generic (PLEG): container finished" podID="b0af1653-4249-4818-b3e2-9d7b3f9aa1f1" containerID="841e56e936369b615e2d88fb773218b05b1c27d7cb4f032a894a1a5f7eee7dcd" exitCode=0 Oct 03 17:50:22 crc kubenswrapper[4744]: I1003 17:50:22.377433 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdjqp" event={"ID":"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1","Type":"ContainerDied","Data":"841e56e936369b615e2d88fb773218b05b1c27d7cb4f032a894a1a5f7eee7dcd"} Oct 03 17:50:22 crc kubenswrapper[4744]: I1003 17:50:22.380355 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 17:50:23 crc kubenswrapper[4744]: I1003 17:50:23.388288 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdjqp" event={"ID":"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1","Type":"ContainerStarted","Data":"9056b5dab353b240f7df315ea42e9269aca7e6ed02f8be739a1214a2becbca7f"} Oct 03 17:50:24 crc kubenswrapper[4744]: I1003 17:50:24.399596 4744 generic.go:334] "Generic (PLEG): container finished" podID="b0af1653-4249-4818-b3e2-9d7b3f9aa1f1" containerID="9056b5dab353b240f7df315ea42e9269aca7e6ed02f8be739a1214a2becbca7f" exitCode=0 Oct 03 17:50:24 crc kubenswrapper[4744]: I1003 17:50:24.399777 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdjqp" event={"ID":"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1","Type":"ContainerDied","Data":"9056b5dab353b240f7df315ea42e9269aca7e6ed02f8be739a1214a2becbca7f"} Oct 03 17:50:25 crc kubenswrapper[4744]: I1003 17:50:25.418425 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdjqp" event={"ID":"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1","Type":"ContainerStarted","Data":"954f45705681fb69ead40b19534e44483665367b967969eb0e33c575cf64629f"} Oct 03 17:50:25 crc kubenswrapper[4744]: I1003 17:50:25.438981 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wdjqp" podStartSLOduration=4.005747956 podStartE2EDuration="6.438965819s" podCreationTimestamp="2025-10-03 17:50:19 +0000 UTC" firstStartedPulling="2025-10-03 17:50:22.380063259 +0000 UTC m=+5148.659939155" lastFinishedPulling="2025-10-03 17:50:24.813281122 +0000 UTC m=+5151.093157018" observedRunningTime="2025-10-03 17:50:25.435035622 +0000 UTC m=+5151.714911528" watchObservedRunningTime="2025-10-03 17:50:25.438965819 +0000 UTC m=+5151.718841715" Oct 03 17:50:30 crc kubenswrapper[4744]: I1003 17:50:30.315678 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:30 crc kubenswrapper[4744]: I1003 17:50:30.316411 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:30 crc kubenswrapper[4744]: I1003 17:50:30.567151 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:30 crc kubenswrapper[4744]: I1003 17:50:30.616439 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:30 crc kubenswrapper[4744]: I1003 17:50:30.816971 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wdjqp"] Oct 03 17:50:32 crc kubenswrapper[4744]: I1003 17:50:32.507748 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wdjqp" podUID="b0af1653-4249-4818-b3e2-9d7b3f9aa1f1" containerName="registry-server" containerID="cri-o://954f45705681fb69ead40b19534e44483665367b967969eb0e33c575cf64629f" gracePeriod=2 Oct 03 17:50:32 crc kubenswrapper[4744]: I1003 17:50:32.962103 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.076175 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-catalog-content\") pod \"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1\" (UID: \"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1\") " Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.076367 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwd2q\" (UniqueName: \"kubernetes.io/projected/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-kube-api-access-gwd2q\") pod \"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1\" (UID: \"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1\") " Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.076474 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-utilities\") pod \"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1\" (UID: \"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1\") " Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.077270 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-utilities" (OuterVolumeSpecName: "utilities") pod "b0af1653-4249-4818-b3e2-9d7b3f9aa1f1" (UID: "b0af1653-4249-4818-b3e2-9d7b3f9aa1f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.081743 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-kube-api-access-gwd2q" (OuterVolumeSpecName: "kube-api-access-gwd2q") pod "b0af1653-4249-4818-b3e2-9d7b3f9aa1f1" (UID: "b0af1653-4249-4818-b3e2-9d7b3f9aa1f1"). InnerVolumeSpecName "kube-api-access-gwd2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.138831 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0af1653-4249-4818-b3e2-9d7b3f9aa1f1" (UID: "b0af1653-4249-4818-b3e2-9d7b3f9aa1f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.178871 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.179117 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwd2q\" (UniqueName: \"kubernetes.io/projected/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-kube-api-access-gwd2q\") on node \"crc\" DevicePath \"\"" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.179196 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.520119 4744 generic.go:334] "Generic (PLEG): container finished" podID="b0af1653-4249-4818-b3e2-9d7b3f9aa1f1" containerID="954f45705681fb69ead40b19534e44483665367b967969eb0e33c575cf64629f" exitCode=0 Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.520194 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdjqp" event={"ID":"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1","Type":"ContainerDied","Data":"954f45705681fb69ead40b19534e44483665367b967969eb0e33c575cf64629f"} Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.520224 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdjqp" event={"ID":"b0af1653-4249-4818-b3e2-9d7b3f9aa1f1","Type":"ContainerDied","Data":"70f8181937c0cd6063d9e695a535adb56bdf256a4d1fd066d5ab1e96fc2d4417"} Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.520241 4744 scope.go:117] "RemoveContainer" containerID="954f45705681fb69ead40b19534e44483665367b967969eb0e33c575cf64629f" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.520634 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdjqp" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.554805 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wdjqp"] Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.564431 4744 scope.go:117] "RemoveContainer" containerID="9056b5dab353b240f7df315ea42e9269aca7e6ed02f8be739a1214a2becbca7f" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.567005 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wdjqp"] Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.592930 4744 scope.go:117] "RemoveContainer" containerID="841e56e936369b615e2d88fb773218b05b1c27d7cb4f032a894a1a5f7eee7dcd" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.632137 4744 scope.go:117] "RemoveContainer" containerID="954f45705681fb69ead40b19534e44483665367b967969eb0e33c575cf64629f" Oct 03 17:50:33 crc kubenswrapper[4744]: E1003 17:50:33.632549 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"954f45705681fb69ead40b19534e44483665367b967969eb0e33c575cf64629f\": container with ID starting with 954f45705681fb69ead40b19534e44483665367b967969eb0e33c575cf64629f not found: ID does not exist" containerID="954f45705681fb69ead40b19534e44483665367b967969eb0e33c575cf64629f" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.632587 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"954f45705681fb69ead40b19534e44483665367b967969eb0e33c575cf64629f"} err="failed to get container status \"954f45705681fb69ead40b19534e44483665367b967969eb0e33c575cf64629f\": rpc error: code = NotFound desc = could not find container \"954f45705681fb69ead40b19534e44483665367b967969eb0e33c575cf64629f\": container with ID starting with 954f45705681fb69ead40b19534e44483665367b967969eb0e33c575cf64629f not found: ID does not exist" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.632613 4744 scope.go:117] "RemoveContainer" containerID="9056b5dab353b240f7df315ea42e9269aca7e6ed02f8be739a1214a2becbca7f" Oct 03 17:50:33 crc kubenswrapper[4744]: E1003 17:50:33.632824 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9056b5dab353b240f7df315ea42e9269aca7e6ed02f8be739a1214a2becbca7f\": container with ID starting with 9056b5dab353b240f7df315ea42e9269aca7e6ed02f8be739a1214a2becbca7f not found: ID does not exist" containerID="9056b5dab353b240f7df315ea42e9269aca7e6ed02f8be739a1214a2becbca7f" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.632854 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9056b5dab353b240f7df315ea42e9269aca7e6ed02f8be739a1214a2becbca7f"} err="failed to get container status \"9056b5dab353b240f7df315ea42e9269aca7e6ed02f8be739a1214a2becbca7f\": rpc error: code = NotFound desc = could not find container \"9056b5dab353b240f7df315ea42e9269aca7e6ed02f8be739a1214a2becbca7f\": container with ID starting with 9056b5dab353b240f7df315ea42e9269aca7e6ed02f8be739a1214a2becbca7f not found: ID does not exist" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.632869 4744 scope.go:117] "RemoveContainer" containerID="841e56e936369b615e2d88fb773218b05b1c27d7cb4f032a894a1a5f7eee7dcd" Oct 03 17:50:33 crc kubenswrapper[4744]: E1003 17:50:33.633062 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"841e56e936369b615e2d88fb773218b05b1c27d7cb4f032a894a1a5f7eee7dcd\": container with ID starting with 841e56e936369b615e2d88fb773218b05b1c27d7cb4f032a894a1a5f7eee7dcd not found: ID does not exist" containerID="841e56e936369b615e2d88fb773218b05b1c27d7cb4f032a894a1a5f7eee7dcd" Oct 03 17:50:33 crc kubenswrapper[4744]: I1003 17:50:33.633088 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"841e56e936369b615e2d88fb773218b05b1c27d7cb4f032a894a1a5f7eee7dcd"} err="failed to get container status \"841e56e936369b615e2d88fb773218b05b1c27d7cb4f032a894a1a5f7eee7dcd\": rpc error: code = NotFound desc = could not find container \"841e56e936369b615e2d88fb773218b05b1c27d7cb4f032a894a1a5f7eee7dcd\": container with ID starting with 841e56e936369b615e2d88fb773218b05b1c27d7cb4f032a894a1a5f7eee7dcd not found: ID does not exist" Oct 03 17:50:34 crc kubenswrapper[4744]: I1003 17:50:34.668570 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:50:34 crc kubenswrapper[4744]: I1003 17:50:34.668930 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:50:34 crc kubenswrapper[4744]: I1003 17:50:34.908426 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0af1653-4249-4818-b3e2-9d7b3f9aa1f1" path="/var/lib/kubelet/pods/b0af1653-4249-4818-b3e2-9d7b3f9aa1f1/volumes" Oct 03 17:50:41 crc kubenswrapper[4744]: I1003 17:50:41.383596 4744 scope.go:117] "RemoveContainer" containerID="058e5cc3ba56f7d11383f600b193dd1611112103fea403dc287147f26667428c" Oct 03 17:50:41 crc kubenswrapper[4744]: I1003 17:50:41.648821 4744 generic.go:334] "Generic (PLEG): container finished" podID="7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8" containerID="7f5e513fdbb52c3368824c4108beb7191433497048df2c33f4b5428f50aa3408" exitCode=0 Oct 03 17:50:41 crc kubenswrapper[4744]: I1003 17:50:41.648959 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgmdg/crc-debug-62tst" event={"ID":"7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8","Type":"ContainerDied","Data":"7f5e513fdbb52c3368824c4108beb7191433497048df2c33f4b5428f50aa3408"} Oct 03 17:50:42 crc kubenswrapper[4744]: I1003 17:50:42.966723 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/crc-debug-62tst" Oct 03 17:50:43 crc kubenswrapper[4744]: I1003 17:50:43.011928 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mgmdg/crc-debug-62tst"] Oct 03 17:50:43 crc kubenswrapper[4744]: I1003 17:50:43.023600 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mgmdg/crc-debug-62tst"] Oct 03 17:50:43 crc kubenswrapper[4744]: I1003 17:50:43.094551 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8-host\") pod \"7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8\" (UID: \"7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8\") " Oct 03 17:50:43 crc kubenswrapper[4744]: I1003 17:50:43.094713 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8-host" (OuterVolumeSpecName: "host") pod "7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8" (UID: "7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 17:50:43 crc kubenswrapper[4744]: I1003 17:50:43.094788 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8fsv\" (UniqueName: \"kubernetes.io/projected/7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8-kube-api-access-q8fsv\") pod \"7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8\" (UID: \"7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8\") " Oct 03 17:50:43 crc kubenswrapper[4744]: I1003 17:50:43.095854 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8-host\") on node \"crc\" DevicePath \"\"" Oct 03 17:50:43 crc kubenswrapper[4744]: I1003 17:50:43.102176 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8-kube-api-access-q8fsv" (OuterVolumeSpecName: "kube-api-access-q8fsv") pod "7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8" (UID: "7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8"). InnerVolumeSpecName "kube-api-access-q8fsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:50:43 crc kubenswrapper[4744]: I1003 17:50:43.198864 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8fsv\" (UniqueName: \"kubernetes.io/projected/7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8-kube-api-access-q8fsv\") on node \"crc\" DevicePath \"\"" Oct 03 17:50:43 crc kubenswrapper[4744]: I1003 17:50:43.676431 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c96e625d8863c983d205e97d9d8eea65ab82994189ee690be72f45096916ac2a" Oct 03 17:50:43 crc kubenswrapper[4744]: I1003 17:50:43.676541 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/crc-debug-62tst" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.216453 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mgmdg/crc-debug-zg9f8"] Oct 03 17:50:44 crc kubenswrapper[4744]: E1003 17:50:44.217216 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8" containerName="container-00" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.217237 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8" containerName="container-00" Oct 03 17:50:44 crc kubenswrapper[4744]: E1003 17:50:44.217265 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0af1653-4249-4818-b3e2-9d7b3f9aa1f1" containerName="extract-utilities" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.217276 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0af1653-4249-4818-b3e2-9d7b3f9aa1f1" containerName="extract-utilities" Oct 03 17:50:44 crc kubenswrapper[4744]: E1003 17:50:44.217307 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0af1653-4249-4818-b3e2-9d7b3f9aa1f1" containerName="extract-content" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.217320 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0af1653-4249-4818-b3e2-9d7b3f9aa1f1" containerName="extract-content" Oct 03 17:50:44 crc kubenswrapper[4744]: E1003 17:50:44.217338 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0af1653-4249-4818-b3e2-9d7b3f9aa1f1" containerName="registry-server" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.217348 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0af1653-4249-4818-b3e2-9d7b3f9aa1f1" containerName="registry-server" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.217695 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8" containerName="container-00" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.217727 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0af1653-4249-4818-b3e2-9d7b3f9aa1f1" containerName="registry-server" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.218587 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/crc-debug-zg9f8" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.326432 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0-host\") pod \"crc-debug-zg9f8\" (UID: \"4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0\") " pod="openshift-must-gather-mgmdg/crc-debug-zg9f8" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.326528 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4kk6\" (UniqueName: \"kubernetes.io/projected/4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0-kube-api-access-k4kk6\") pod \"crc-debug-zg9f8\" (UID: \"4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0\") " pod="openshift-must-gather-mgmdg/crc-debug-zg9f8" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.428384 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0-host\") pod \"crc-debug-zg9f8\" (UID: \"4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0\") " pod="openshift-must-gather-mgmdg/crc-debug-zg9f8" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.428753 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4kk6\" (UniqueName: \"kubernetes.io/projected/4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0-kube-api-access-k4kk6\") pod \"crc-debug-zg9f8\" (UID: \"4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0\") " pod="openshift-must-gather-mgmdg/crc-debug-zg9f8" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.428583 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0-host\") pod \"crc-debug-zg9f8\" (UID: \"4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0\") " pod="openshift-must-gather-mgmdg/crc-debug-zg9f8" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.452157 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4kk6\" (UniqueName: \"kubernetes.io/projected/4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0-kube-api-access-k4kk6\") pod \"crc-debug-zg9f8\" (UID: \"4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0\") " pod="openshift-must-gather-mgmdg/crc-debug-zg9f8" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.549025 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/crc-debug-zg9f8" Oct 03 17:50:44 crc kubenswrapper[4744]: I1003 17:50:44.909804 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8" path="/var/lib/kubelet/pods/7ae3990a-bbfe-4dcc-9c64-e1c7dec769a8/volumes" Oct 03 17:50:45 crc kubenswrapper[4744]: I1003 17:50:45.701277 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgmdg/crc-debug-zg9f8" event={"ID":"4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0","Type":"ContainerStarted","Data":"8e8353427571eb1dc247ddf819b45b99a264d12bd7a20d17716a1adf33feba98"} Oct 03 17:50:45 crc kubenswrapper[4744]: I1003 17:50:45.701707 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgmdg/crc-debug-zg9f8" event={"ID":"4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0","Type":"ContainerStarted","Data":"cf6ca94a6f2b1ffbccfab78c44d49e3b9279d3c2c252b391e347146393025f13"} Oct 03 17:50:45 crc kubenswrapper[4744]: I1003 17:50:45.722736 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mgmdg/crc-debug-zg9f8" podStartSLOduration=1.7227078 podStartE2EDuration="1.7227078s" podCreationTimestamp="2025-10-03 17:50:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 17:50:45.719066219 +0000 UTC m=+5171.998942135" watchObservedRunningTime="2025-10-03 17:50:45.7227078 +0000 UTC m=+5172.002583736" Oct 03 17:50:46 crc kubenswrapper[4744]: I1003 17:50:46.712412 4744 generic.go:334] "Generic (PLEG): container finished" podID="4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0" containerID="8e8353427571eb1dc247ddf819b45b99a264d12bd7a20d17716a1adf33feba98" exitCode=0 Oct 03 17:50:46 crc kubenswrapper[4744]: I1003 17:50:46.712505 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgmdg/crc-debug-zg9f8" event={"ID":"4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0","Type":"ContainerDied","Data":"8e8353427571eb1dc247ddf819b45b99a264d12bd7a20d17716a1adf33feba98"} Oct 03 17:50:47 crc kubenswrapper[4744]: I1003 17:50:47.843191 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/crc-debug-zg9f8" Oct 03 17:50:48 crc kubenswrapper[4744]: I1003 17:50:48.000619 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0-host\") pod \"4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0\" (UID: \"4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0\") " Oct 03 17:50:48 crc kubenswrapper[4744]: I1003 17:50:48.000728 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4kk6\" (UniqueName: \"kubernetes.io/projected/4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0-kube-api-access-k4kk6\") pod \"4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0\" (UID: \"4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0\") " Oct 03 17:50:48 crc kubenswrapper[4744]: I1003 17:50:48.001754 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0-host" (OuterVolumeSpecName: "host") pod "4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0" (UID: "4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 17:50:48 crc kubenswrapper[4744]: I1003 17:50:48.006659 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0-kube-api-access-k4kk6" (OuterVolumeSpecName: "kube-api-access-k4kk6") pod "4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0" (UID: "4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0"). InnerVolumeSpecName "kube-api-access-k4kk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:50:48 crc kubenswrapper[4744]: I1003 17:50:48.102378 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0-host\") on node \"crc\" DevicePath \"\"" Oct 03 17:50:48 crc kubenswrapper[4744]: I1003 17:50:48.102406 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4kk6\" (UniqueName: \"kubernetes.io/projected/4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0-kube-api-access-k4kk6\") on node \"crc\" DevicePath \"\"" Oct 03 17:50:48 crc kubenswrapper[4744]: I1003 17:50:48.751741 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgmdg/crc-debug-zg9f8" event={"ID":"4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0","Type":"ContainerDied","Data":"cf6ca94a6f2b1ffbccfab78c44d49e3b9279d3c2c252b391e347146393025f13"} Oct 03 17:50:48 crc kubenswrapper[4744]: I1003 17:50:48.752051 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf6ca94a6f2b1ffbccfab78c44d49e3b9279d3c2c252b391e347146393025f13" Oct 03 17:50:48 crc kubenswrapper[4744]: I1003 17:50:48.752099 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/crc-debug-zg9f8" Oct 03 17:50:55 crc kubenswrapper[4744]: I1003 17:50:55.086602 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mgmdg/crc-debug-zg9f8"] Oct 03 17:50:55 crc kubenswrapper[4744]: I1003 17:50:55.093992 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mgmdg/crc-debug-zg9f8"] Oct 03 17:50:56 crc kubenswrapper[4744]: I1003 17:50:56.289716 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mgmdg/crc-debug-bqxgx"] Oct 03 17:50:56 crc kubenswrapper[4744]: E1003 17:50:56.290429 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0" containerName="container-00" Oct 03 17:50:56 crc kubenswrapper[4744]: I1003 17:50:56.290443 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0" containerName="container-00" Oct 03 17:50:56 crc kubenswrapper[4744]: I1003 17:50:56.290701 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0" containerName="container-00" Oct 03 17:50:56 crc kubenswrapper[4744]: I1003 17:50:56.291442 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/crc-debug-bqxgx" Oct 03 17:50:56 crc kubenswrapper[4744]: I1003 17:50:56.456561 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/270443ae-cab5-4bae-848f-76bcdb3fb610-host\") pod \"crc-debug-bqxgx\" (UID: \"270443ae-cab5-4bae-848f-76bcdb3fb610\") " pod="openshift-must-gather-mgmdg/crc-debug-bqxgx" Oct 03 17:50:56 crc kubenswrapper[4744]: I1003 17:50:56.457060 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkd5p\" (UniqueName: \"kubernetes.io/projected/270443ae-cab5-4bae-848f-76bcdb3fb610-kube-api-access-vkd5p\") pod \"crc-debug-bqxgx\" (UID: \"270443ae-cab5-4bae-848f-76bcdb3fb610\") " pod="openshift-must-gather-mgmdg/crc-debug-bqxgx" Oct 03 17:50:56 crc kubenswrapper[4744]: I1003 17:50:56.559866 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkd5p\" (UniqueName: \"kubernetes.io/projected/270443ae-cab5-4bae-848f-76bcdb3fb610-kube-api-access-vkd5p\") pod \"crc-debug-bqxgx\" (UID: \"270443ae-cab5-4bae-848f-76bcdb3fb610\") " pod="openshift-must-gather-mgmdg/crc-debug-bqxgx" Oct 03 17:50:56 crc kubenswrapper[4744]: I1003 17:50:56.560138 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/270443ae-cab5-4bae-848f-76bcdb3fb610-host\") pod \"crc-debug-bqxgx\" (UID: \"270443ae-cab5-4bae-848f-76bcdb3fb610\") " pod="openshift-must-gather-mgmdg/crc-debug-bqxgx" Oct 03 17:50:56 crc kubenswrapper[4744]: I1003 17:50:56.560341 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/270443ae-cab5-4bae-848f-76bcdb3fb610-host\") pod \"crc-debug-bqxgx\" (UID: \"270443ae-cab5-4bae-848f-76bcdb3fb610\") " pod="openshift-must-gather-mgmdg/crc-debug-bqxgx" Oct 03 17:50:56 crc kubenswrapper[4744]: I1003 17:50:56.594862 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkd5p\" (UniqueName: \"kubernetes.io/projected/270443ae-cab5-4bae-848f-76bcdb3fb610-kube-api-access-vkd5p\") pod \"crc-debug-bqxgx\" (UID: \"270443ae-cab5-4bae-848f-76bcdb3fb610\") " pod="openshift-must-gather-mgmdg/crc-debug-bqxgx" Oct 03 17:50:56 crc kubenswrapper[4744]: I1003 17:50:56.613058 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/crc-debug-bqxgx" Oct 03 17:50:56 crc kubenswrapper[4744]: I1003 17:50:56.833625 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgmdg/crc-debug-bqxgx" event={"ID":"270443ae-cab5-4bae-848f-76bcdb3fb610","Type":"ContainerStarted","Data":"1dc14471efa266b3035f20e4eac36a09ca776ed405252e78d01b6e06d10c3bba"} Oct 03 17:50:56 crc kubenswrapper[4744]: I1003 17:50:56.911168 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0" path="/var/lib/kubelet/pods/4f7d387a-a577-4a8e-a9e8-cdcf4ffb28c0/volumes" Oct 03 17:50:57 crc kubenswrapper[4744]: I1003 17:50:57.848398 4744 generic.go:334] "Generic (PLEG): container finished" podID="270443ae-cab5-4bae-848f-76bcdb3fb610" containerID="8a2cdd87d02dbfcd3ee5f29e3326f58c92c6320b07c6e8fb1d9fcacf6b321094" exitCode=0 Oct 03 17:50:57 crc kubenswrapper[4744]: I1003 17:50:57.848813 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgmdg/crc-debug-bqxgx" event={"ID":"270443ae-cab5-4bae-848f-76bcdb3fb610","Type":"ContainerDied","Data":"8a2cdd87d02dbfcd3ee5f29e3326f58c92c6320b07c6e8fb1d9fcacf6b321094"} Oct 03 17:50:57 crc kubenswrapper[4744]: I1003 17:50:57.901314 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mgmdg/crc-debug-bqxgx"] Oct 03 17:50:57 crc kubenswrapper[4744]: I1003 17:50:57.912333 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mgmdg/crc-debug-bqxgx"] Oct 03 17:50:58 crc kubenswrapper[4744]: I1003 17:50:58.967880 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/crc-debug-bqxgx" Oct 03 17:50:59 crc kubenswrapper[4744]: I1003 17:50:59.117689 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkd5p\" (UniqueName: \"kubernetes.io/projected/270443ae-cab5-4bae-848f-76bcdb3fb610-kube-api-access-vkd5p\") pod \"270443ae-cab5-4bae-848f-76bcdb3fb610\" (UID: \"270443ae-cab5-4bae-848f-76bcdb3fb610\") " Oct 03 17:50:59 crc kubenswrapper[4744]: I1003 17:50:59.118001 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/270443ae-cab5-4bae-848f-76bcdb3fb610-host\") pod \"270443ae-cab5-4bae-848f-76bcdb3fb610\" (UID: \"270443ae-cab5-4bae-848f-76bcdb3fb610\") " Oct 03 17:50:59 crc kubenswrapper[4744]: I1003 17:50:59.118147 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270443ae-cab5-4bae-848f-76bcdb3fb610-host" (OuterVolumeSpecName: "host") pod "270443ae-cab5-4bae-848f-76bcdb3fb610" (UID: "270443ae-cab5-4bae-848f-76bcdb3fb610"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 17:50:59 crc kubenswrapper[4744]: I1003 17:50:59.118870 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/270443ae-cab5-4bae-848f-76bcdb3fb610-host\") on node \"crc\" DevicePath \"\"" Oct 03 17:50:59 crc kubenswrapper[4744]: I1003 17:50:59.123834 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/270443ae-cab5-4bae-848f-76bcdb3fb610-kube-api-access-vkd5p" (OuterVolumeSpecName: "kube-api-access-vkd5p") pod "270443ae-cab5-4bae-848f-76bcdb3fb610" (UID: "270443ae-cab5-4bae-848f-76bcdb3fb610"). InnerVolumeSpecName "kube-api-access-vkd5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:50:59 crc kubenswrapper[4744]: I1003 17:50:59.220481 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkd5p\" (UniqueName: \"kubernetes.io/projected/270443ae-cab5-4bae-848f-76bcdb3fb610-kube-api-access-vkd5p\") on node \"crc\" DevicePath \"\"" Oct 03 17:50:59 crc kubenswrapper[4744]: I1003 17:50:59.568150 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx_aaa2ef65-fb80-4f95-81fc-29841e8b6847/util/0.log" Oct 03 17:50:59 crc kubenswrapper[4744]: I1003 17:50:59.723728 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx_aaa2ef65-fb80-4f95-81fc-29841e8b6847/pull/0.log" Oct 03 17:50:59 crc kubenswrapper[4744]: I1003 17:50:59.736114 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx_aaa2ef65-fb80-4f95-81fc-29841e8b6847/util/0.log" Oct 03 17:50:59 crc kubenswrapper[4744]: I1003 17:50:59.805361 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx_aaa2ef65-fb80-4f95-81fc-29841e8b6847/pull/0.log" Oct 03 17:50:59 crc kubenswrapper[4744]: I1003 17:50:59.867543 4744 scope.go:117] "RemoveContainer" containerID="8a2cdd87d02dbfcd3ee5f29e3326f58c92c6320b07c6e8fb1d9fcacf6b321094" Oct 03 17:50:59 crc kubenswrapper[4744]: I1003 17:50:59.867600 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/crc-debug-bqxgx" Oct 03 17:50:59 crc kubenswrapper[4744]: I1003 17:50:59.979170 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx_aaa2ef65-fb80-4f95-81fc-29841e8b6847/extract/0.log" Oct 03 17:51:00 crc kubenswrapper[4744]: I1003 17:51:00.004939 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx_aaa2ef65-fb80-4f95-81fc-29841e8b6847/pull/0.log" Oct 03 17:51:00 crc kubenswrapper[4744]: I1003 17:51:00.007208 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_43e4cf20d1fc0e46d9f23a6c11fa46fd07b505ffdc2dc6d3ce54e4831bgr2hx_aaa2ef65-fb80-4f95-81fc-29841e8b6847/util/0.log" Oct 03 17:51:00 crc kubenswrapper[4744]: I1003 17:51:00.149089 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6c675fb79f-4st8l_d68c3557-6655-4449-8167-c28d1ccf442e/kube-rbac-proxy/0.log" Oct 03 17:51:00 crc kubenswrapper[4744]: I1003 17:51:00.214213 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79d68d6c85-4r792_829ba77b-d3cc-416c-8813-eeff7b6b60bd/kube-rbac-proxy/0.log" Oct 03 17:51:00 crc kubenswrapper[4744]: I1003 17:51:00.230111 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6c675fb79f-4st8l_d68c3557-6655-4449-8167-c28d1ccf442e/manager/0.log" Oct 03 17:51:00 crc kubenswrapper[4744]: I1003 17:51:00.399889 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-bs7f6_19371562-3c8d-4dce-864c-0a37cbce2bf3/kube-rbac-proxy/0.log" Oct 03 17:51:00 crc kubenswrapper[4744]: I1003 17:51:00.405115 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79d68d6c85-4r792_829ba77b-d3cc-416c-8813-eeff7b6b60bd/manager/0.log" Oct 03 17:51:00 crc kubenswrapper[4744]: I1003 17:51:00.442159 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-bs7f6_19371562-3c8d-4dce-864c-0a37cbce2bf3/manager/0.log" Oct 03 17:51:00 crc kubenswrapper[4744]: I1003 17:51:00.637863 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-846dff85b5-7qtkn_a38be30f-e956-4859-9610-a4d0735fe9d4/kube-rbac-proxy/0.log" Oct 03 17:51:00 crc kubenswrapper[4744]: I1003 17:51:00.715351 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-846dff85b5-7qtkn_a38be30f-e956-4859-9610-a4d0735fe9d4/manager/0.log" Oct 03 17:51:00 crc kubenswrapper[4744]: I1003 17:51:00.843306 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-599898f689-fxkn9_da0c0550-089f-4bd9-aaf8-960fcc75f019/manager/0.log" Oct 03 17:51:00 crc kubenswrapper[4744]: I1003 17:51:00.846295 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-599898f689-fxkn9_da0c0550-089f-4bd9-aaf8-960fcc75f019/kube-rbac-proxy/0.log" Oct 03 17:51:00 crc kubenswrapper[4744]: I1003 17:51:00.901068 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="270443ae-cab5-4bae-848f-76bcdb3fb610" path="/var/lib/kubelet/pods/270443ae-cab5-4bae-848f-76bcdb3fb610/volumes" Oct 03 17:51:00 crc kubenswrapper[4744]: I1003 17:51:00.922764 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6769b867d9-n26jb_fd8ecb24-30b8-4012-846a-4618f178996b/kube-rbac-proxy/0.log" Oct 03 17:51:01 crc kubenswrapper[4744]: I1003 17:51:01.017377 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6769b867d9-n26jb_fd8ecb24-30b8-4012-846a-4618f178996b/manager/0.log" Oct 03 17:51:01 crc kubenswrapper[4744]: I1003 17:51:01.135926 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5fbf469cd7-4jdmt_3bc36b86-4888-4341-b52d-b9b73013f9ad/kube-rbac-proxy/0.log" Oct 03 17:51:01 crc kubenswrapper[4744]: I1003 17:51:01.227582 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5fbf469cd7-4jdmt_3bc36b86-4888-4341-b52d-b9b73013f9ad/manager/0.log" Oct 03 17:51:01 crc kubenswrapper[4744]: I1003 17:51:01.294519 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-84bc9db6cc-tb5k2_23519c20-6aec-4461-bea9-adaf88e87ab3/kube-rbac-proxy/0.log" Oct 03 17:51:01 crc kubenswrapper[4744]: I1003 17:51:01.338912 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-84bc9db6cc-tb5k2_23519c20-6aec-4461-bea9-adaf88e87ab3/manager/0.log" Oct 03 17:51:01 crc kubenswrapper[4744]: I1003 17:51:01.397478 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7f55849f88-r8tdt_a862283f-4243-42bd-bb2c-5b2392d657ac/kube-rbac-proxy/0.log" Oct 03 17:51:01 crc kubenswrapper[4744]: I1003 17:51:01.504297 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7f55849f88-r8tdt_a862283f-4243-42bd-bb2c-5b2392d657ac/manager/0.log" Oct 03 17:51:01 crc kubenswrapper[4744]: I1003 17:51:01.618361 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6fd6854b49-zd5fc_be9443a6-46c8-4374-8caa-8040ae117981/kube-rbac-proxy/0.log" Oct 03 17:51:01 crc kubenswrapper[4744]: I1003 17:51:01.644033 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6fd6854b49-zd5fc_be9443a6-46c8-4374-8caa-8040ae117981/manager/0.log" Oct 03 17:51:01 crc kubenswrapper[4744]: I1003 17:51:01.792277 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5c468bf4d4-kzsg2_d34995fe-607b-46ae-a2eb-c91d9c244415/kube-rbac-proxy/0.log" Oct 03 17:51:01 crc kubenswrapper[4744]: I1003 17:51:01.837659 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5c468bf4d4-kzsg2_d34995fe-607b-46ae-a2eb-c91d9c244415/manager/0.log" Oct 03 17:51:01 crc kubenswrapper[4744]: I1003 17:51:01.923023 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6574bf987d-xxdpr_2832b841-b383-4eed-a8ce-71062269238a/kube-rbac-proxy/0.log" Oct 03 17:51:01 crc kubenswrapper[4744]: I1003 17:51:01.990069 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6574bf987d-xxdpr_2832b841-b383-4eed-a8ce-71062269238a/manager/0.log" Oct 03 17:51:02 crc kubenswrapper[4744]: I1003 17:51:02.024548 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-555c7456bd-6xg8v_438b2904-4b85-4d82-b883-f5788d4caab9/kube-rbac-proxy/0.log" Oct 03 17:51:02 crc kubenswrapper[4744]: I1003 17:51:02.202090 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-59d6cfdf45-xr6db_d34377f2-1b67-4c62-9602-ddb93f0cb4d7/manager/0.log" Oct 03 17:51:02 crc kubenswrapper[4744]: I1003 17:51:02.214952 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-59d6cfdf45-xr6db_d34377f2-1b67-4c62-9602-ddb93f0cb4d7/kube-rbac-proxy/0.log" Oct 03 17:51:02 crc kubenswrapper[4744]: I1003 17:51:02.220392 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-555c7456bd-6xg8v_438b2904-4b85-4d82-b883-f5788d4caab9/manager/0.log" Oct 03 17:51:02 crc kubenswrapper[4744]: I1003 17:51:02.354947 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6f64c4d678v879q_911d5ffd-d447-4b8a-a177-7fc738d2aa35/kube-rbac-proxy/0.log" Oct 03 17:51:02 crc kubenswrapper[4744]: I1003 17:51:02.386478 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6f64c4d678v879q_911d5ffd-d447-4b8a-a177-7fc738d2aa35/manager/0.log" Oct 03 17:51:02 crc kubenswrapper[4744]: I1003 17:51:02.532837 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7758d4b55d-6xczl_4717fade-30fb-4f78-b7fb-dec2799d9b8c/kube-rbac-proxy/0.log" Oct 03 17:51:02 crc kubenswrapper[4744]: I1003 17:51:02.576300 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-d94d76459-krksv_df419359-3ee4-4c00-bb37-c7387b19492a/kube-rbac-proxy/0.log" Oct 03 17:51:02 crc kubenswrapper[4744]: I1003 17:51:02.801960 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-n2bfv_d8b895bc-8e19-4b92-999f-d579c1d51e51/registry-server/0.log" Oct 03 17:51:02 crc kubenswrapper[4744]: I1003 17:51:02.885877 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-d94d76459-krksv_df419359-3ee4-4c00-bb37-c7387b19492a/operator/0.log" Oct 03 17:51:03 crc kubenswrapper[4744]: I1003 17:51:03.014338 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-688db7b6c7-jhqdv_8be63e93-5f9d-4ffe-9ab2-d5b794c34f9b/kube-rbac-proxy/0.log" Oct 03 17:51:03 crc kubenswrapper[4744]: I1003 17:51:03.140779 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-7d8bb7f44c-jktv9_39de7155-617f-444a-bb16-0cb7fb229eb7/kube-rbac-proxy/0.log" Oct 03 17:51:03 crc kubenswrapper[4744]: I1003 17:51:03.172867 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-688db7b6c7-jhqdv_8be63e93-5f9d-4ffe-9ab2-d5b794c34f9b/manager/0.log" Oct 03 17:51:03 crc kubenswrapper[4744]: I1003 17:51:03.299227 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-7d8bb7f44c-jktv9_39de7155-617f-444a-bb16-0cb7fb229eb7/manager/0.log" Oct 03 17:51:03 crc kubenswrapper[4744]: I1003 17:51:03.437449 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-2g6tw_aef71e84-6035-42fb-a16f-c0e97876f1c0/operator/0.log" Oct 03 17:51:03 crc kubenswrapper[4744]: I1003 17:51:03.545615 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-psqc4_a6943fb4-ab7a-445f-b4d1-be9089562b77/kube-rbac-proxy/0.log" Oct 03 17:51:03 crc kubenswrapper[4744]: I1003 17:51:03.576479 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-psqc4_a6943fb4-ab7a-445f-b4d1-be9089562b77/manager/0.log" Oct 03 17:51:03 crc kubenswrapper[4744]: I1003 17:51:03.648862 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5db5cf686f-7qgsm_09c1f75f-d699-4675-b78d-a37353acd758/kube-rbac-proxy/0.log" Oct 03 17:51:03 crc kubenswrapper[4744]: I1003 17:51:03.726587 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7758d4b55d-6xczl_4717fade-30fb-4f78-b7fb-dec2799d9b8c/manager/0.log" Oct 03 17:51:03 crc kubenswrapper[4744]: I1003 17:51:03.786615 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5db5cf686f-7qgsm_09c1f75f-d699-4675-b78d-a37353acd758/manager/0.log" Oct 03 17:51:04 crc kubenswrapper[4744]: I1003 17:51:04.186238 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-ksw89_1d753e81-b04a-4216-991f-256a2ab6ac94/manager/0.log" Oct 03 17:51:04 crc kubenswrapper[4744]: I1003 17:51:04.231611 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-fcd7d9895-wjl4s_1aa4d11e-aebd-477b-98e1-a05ea74a7677/kube-rbac-proxy/0.log" Oct 03 17:51:04 crc kubenswrapper[4744]: I1003 17:51:04.236949 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-fcd7d9895-wjl4s_1aa4d11e-aebd-477b-98e1-a05ea74a7677/manager/0.log" Oct 03 17:51:04 crc kubenswrapper[4744]: I1003 17:51:04.251755 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-ksw89_1d753e81-b04a-4216-991f-256a2ab6ac94/kube-rbac-proxy/0.log" Oct 03 17:51:04 crc kubenswrapper[4744]: I1003 17:51:04.668340 4744 patch_prober.go:28] interesting pod/machine-config-daemon-qzw8f container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 17:51:04 crc kubenswrapper[4744]: I1003 17:51:04.668409 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 17:51:04 crc kubenswrapper[4744]: I1003 17:51:04.668460 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" Oct 03 17:51:04 crc kubenswrapper[4744]: I1003 17:51:04.669092 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2"} pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 17:51:04 crc kubenswrapper[4744]: I1003 17:51:04.669167 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerName="machine-config-daemon" containerID="cri-o://8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" gracePeriod=600 Oct 03 17:51:04 crc kubenswrapper[4744]: E1003 17:51:04.793271 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:51:04 crc kubenswrapper[4744]: I1003 17:51:04.922063 4744 generic.go:334] "Generic (PLEG): container finished" podID="cc9bea96-4847-44c8-ab50-d469a6c6155d" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" exitCode=0 Oct 03 17:51:04 crc kubenswrapper[4744]: I1003 17:51:04.922109 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerDied","Data":"8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2"} Oct 03 17:51:04 crc kubenswrapper[4744]: I1003 17:51:04.922140 4744 scope.go:117] "RemoveContainer" containerID="c1441960a66e4ddfd471f97e8bfbe584c3f6a01e4d0501c4d81cf0e95cf9e35e" Oct 03 17:51:04 crc kubenswrapper[4744]: I1003 17:51:04.924212 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:51:04 crc kubenswrapper[4744]: E1003 17:51:04.924824 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:51:15 crc kubenswrapper[4744]: I1003 17:51:15.892665 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:51:15 crc kubenswrapper[4744]: E1003 17:51:15.893520 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:51:20 crc kubenswrapper[4744]: I1003 17:51:20.774039 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-qmzm9_c41f7473-52d9-4d32-9364-ed61efb65af4/control-plane-machine-set-operator/0.log" Oct 03 17:51:20 crc kubenswrapper[4744]: I1003 17:51:20.939683 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-btrfh_23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c/kube-rbac-proxy/0.log" Oct 03 17:51:20 crc kubenswrapper[4744]: I1003 17:51:20.983459 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-btrfh_23b0d312-c1eb-47f2-a1ae-7fbcbc5ebb8c/machine-api-operator/0.log" Oct 03 17:51:29 crc kubenswrapper[4744]: I1003 17:51:29.891693 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:51:29 crc kubenswrapper[4744]: E1003 17:51:29.892664 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:51:34 crc kubenswrapper[4744]: I1003 17:51:34.815350 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-m8bhx_15fdf250-5959-4502-b7be-4d7af4eff05d/cert-manager-controller/0.log" Oct 03 17:51:35 crc kubenswrapper[4744]: I1003 17:51:35.055732 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-xjmdh_f62bcdb5-0b26-4be1-af80-c413f76ce50c/cert-manager-webhook/0.log" Oct 03 17:51:35 crc kubenswrapper[4744]: I1003 17:51:35.058653 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-xp59p_862c46f4-2e50-4556-8405-bfcdb337ed95/cert-manager-cainjector/0.log" Oct 03 17:51:42 crc kubenswrapper[4744]: I1003 17:51:42.893764 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:51:42 crc kubenswrapper[4744]: E1003 17:51:42.894791 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:51:48 crc kubenswrapper[4744]: I1003 17:51:48.393129 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-587v7_c5f8637e-bf33-4533-970d-a21b8837ce78/nmstate-console-plugin/0.log" Oct 03 17:51:48 crc kubenswrapper[4744]: I1003 17:51:48.730768 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-cvn5h_d0fe8f50-20f3-483f-8904-8f084fba1fb1/nmstate-handler/0.log" Oct 03 17:51:48 crc kubenswrapper[4744]: I1003 17:51:48.804768 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-pdpss_4039e497-2af2-483a-8a21-14f14df6df33/nmstate-metrics/0.log" Oct 03 17:51:48 crc kubenswrapper[4744]: I1003 17:51:48.805118 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-pdpss_4039e497-2af2-483a-8a21-14f14df6df33/kube-rbac-proxy/0.log" Oct 03 17:51:48 crc kubenswrapper[4744]: I1003 17:51:48.986626 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-dmz4f_63038935-53b5-4371-ab75-80e6c1522980/nmstate-webhook/0.log" Oct 03 17:51:49 crc kubenswrapper[4744]: I1003 17:51:49.020706 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-qd582_af44daee-90fc-4ccd-a4dd-ae2513f0fe2c/nmstate-operator/0.log" Oct 03 17:51:56 crc kubenswrapper[4744]: I1003 17:51:56.892491 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:51:56 crc kubenswrapper[4744]: E1003 17:51:56.893186 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:52:05 crc kubenswrapper[4744]: I1003 17:52:05.546962 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-vhk6s_5f04d600-8bb7-4420-aa5f-c1b51a0a4e16/kube-rbac-proxy/0.log" Oct 03 17:52:05 crc kubenswrapper[4744]: I1003 17:52:05.618313 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-vhk6s_5f04d600-8bb7-4420-aa5f-c1b51a0a4e16/controller/0.log" Oct 03 17:52:05 crc kubenswrapper[4744]: I1003 17:52:05.678796 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-frr-files/0.log" Oct 03 17:52:05 crc kubenswrapper[4744]: I1003 17:52:05.865844 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-frr-files/0.log" Oct 03 17:52:05 crc kubenswrapper[4744]: I1003 17:52:05.870703 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-reloader/0.log" Oct 03 17:52:05 crc kubenswrapper[4744]: I1003 17:52:05.904442 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-metrics/0.log" Oct 03 17:52:05 crc kubenswrapper[4744]: I1003 17:52:05.906591 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-reloader/0.log" Oct 03 17:52:06 crc kubenswrapper[4744]: I1003 17:52:06.100109 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-frr-files/0.log" Oct 03 17:52:06 crc kubenswrapper[4744]: I1003 17:52:06.106233 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-metrics/0.log" Oct 03 17:52:06 crc kubenswrapper[4744]: I1003 17:52:06.116908 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-metrics/0.log" Oct 03 17:52:06 crc kubenswrapper[4744]: I1003 17:52:06.159244 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-reloader/0.log" Oct 03 17:52:06 crc kubenswrapper[4744]: I1003 17:52:06.312240 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-metrics/0.log" Oct 03 17:52:06 crc kubenswrapper[4744]: I1003 17:52:06.321295 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-reloader/0.log" Oct 03 17:52:06 crc kubenswrapper[4744]: I1003 17:52:06.323779 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/cp-frr-files/0.log" Oct 03 17:52:06 crc kubenswrapper[4744]: I1003 17:52:06.344881 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/controller/0.log" Oct 03 17:52:06 crc kubenswrapper[4744]: I1003 17:52:06.507955 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/kube-rbac-proxy/0.log" Oct 03 17:52:06 crc kubenswrapper[4744]: I1003 17:52:06.515527 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/kube-rbac-proxy-frr/0.log" Oct 03 17:52:06 crc kubenswrapper[4744]: I1003 17:52:06.545926 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/frr-metrics/0.log" Oct 03 17:52:06 crc kubenswrapper[4744]: I1003 17:52:06.745437 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/reloader/0.log" Oct 03 17:52:06 crc kubenswrapper[4744]: I1003 17:52:06.797125 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-xsl5f_703cac15-6674-4529-9a6b-75f646a84d3e/frr-k8s-webhook-server/0.log" Oct 03 17:52:06 crc kubenswrapper[4744]: I1003 17:52:06.956595 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6cbff679b9-cdsl8_8776a23e-e0c0-4c8d-84e8-d34d5036e3c6/manager/0.log" Oct 03 17:52:07 crc kubenswrapper[4744]: I1003 17:52:07.218588 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-86c9f64ddf-r6kfg_259f45bb-e4ab-45f4-9d77-5e788361c35a/webhook-server/0.log" Oct 03 17:52:07 crc kubenswrapper[4744]: I1003 17:52:07.270566 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5qx66_0d474bc7-3b94-40d8-91d6-db976b2975d7/kube-rbac-proxy/0.log" Oct 03 17:52:07 crc kubenswrapper[4744]: I1003 17:52:07.829483 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-5qx66_0d474bc7-3b94-40d8-91d6-db976b2975d7/speaker/0.log" Oct 03 17:52:07 crc kubenswrapper[4744]: I1003 17:52:07.891522 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:52:07 crc kubenswrapper[4744]: E1003 17:52:07.891804 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:52:07 crc kubenswrapper[4744]: I1003 17:52:07.994324 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7dg9g_56173127-d55d-4c05-ac7b-1cdc4484f3bb/frr/0.log" Oct 03 17:52:21 crc kubenswrapper[4744]: I1003 17:52:21.892436 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:52:21 crc kubenswrapper[4744]: E1003 17:52:21.893458 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:52:22 crc kubenswrapper[4744]: I1003 17:52:22.080777 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh_c795770c-99b7-4077-a8b4-f50b678f32b8/util/0.log" Oct 03 17:52:22 crc kubenswrapper[4744]: I1003 17:52:22.194867 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh_c795770c-99b7-4077-a8b4-f50b678f32b8/pull/0.log" Oct 03 17:52:22 crc kubenswrapper[4744]: I1003 17:52:22.214933 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh_c795770c-99b7-4077-a8b4-f50b678f32b8/util/0.log" Oct 03 17:52:22 crc kubenswrapper[4744]: I1003 17:52:22.268378 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh_c795770c-99b7-4077-a8b4-f50b678f32b8/pull/0.log" Oct 03 17:52:22 crc kubenswrapper[4744]: I1003 17:52:22.464616 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh_c795770c-99b7-4077-a8b4-f50b678f32b8/extract/0.log" Oct 03 17:52:22 crc kubenswrapper[4744]: I1003 17:52:22.472762 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh_c795770c-99b7-4077-a8b4-f50b678f32b8/pull/0.log" Oct 03 17:52:22 crc kubenswrapper[4744]: I1003 17:52:22.479414 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d25jbgh_c795770c-99b7-4077-a8b4-f50b678f32b8/util/0.log" Oct 03 17:52:22 crc kubenswrapper[4744]: I1003 17:52:22.627799 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcfbk_f6fb66df-a4db-417b-90fc-c779777883be/extract-utilities/0.log" Oct 03 17:52:22 crc kubenswrapper[4744]: I1003 17:52:22.754906 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcfbk_f6fb66df-a4db-417b-90fc-c779777883be/extract-content/0.log" Oct 03 17:52:22 crc kubenswrapper[4744]: I1003 17:52:22.808210 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcfbk_f6fb66df-a4db-417b-90fc-c779777883be/extract-utilities/0.log" Oct 03 17:52:22 crc kubenswrapper[4744]: I1003 17:52:22.818800 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcfbk_f6fb66df-a4db-417b-90fc-c779777883be/extract-content/0.log" Oct 03 17:52:22 crc kubenswrapper[4744]: I1003 17:52:22.988335 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcfbk_f6fb66df-a4db-417b-90fc-c779777883be/extract-utilities/0.log" Oct 03 17:52:22 crc kubenswrapper[4744]: I1003 17:52:22.988771 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcfbk_f6fb66df-a4db-417b-90fc-c779777883be/extract-content/0.log" Oct 03 17:52:23 crc kubenswrapper[4744]: I1003 17:52:23.198176 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7v9wq_6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2/extract-utilities/0.log" Oct 03 17:52:23 crc kubenswrapper[4744]: I1003 17:52:23.408266 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7v9wq_6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2/extract-content/0.log" Oct 03 17:52:23 crc kubenswrapper[4744]: I1003 17:52:23.425208 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7v9wq_6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2/extract-utilities/0.log" Oct 03 17:52:23 crc kubenswrapper[4744]: I1003 17:52:23.511531 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7v9wq_6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2/extract-content/0.log" Oct 03 17:52:23 crc kubenswrapper[4744]: I1003 17:52:23.519244 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vcfbk_f6fb66df-a4db-417b-90fc-c779777883be/registry-server/0.log" Oct 03 17:52:23 crc kubenswrapper[4744]: I1003 17:52:23.700786 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7v9wq_6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2/extract-content/0.log" Oct 03 17:52:23 crc kubenswrapper[4744]: I1003 17:52:23.700799 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7v9wq_6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2/extract-utilities/0.log" Oct 03 17:52:23 crc kubenswrapper[4744]: I1003 17:52:23.917966 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd_6a7d3bdf-d461-4fcb-810c-421130e8d524/util/0.log" Oct 03 17:52:24 crc kubenswrapper[4744]: I1003 17:52:24.124300 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd_6a7d3bdf-d461-4fcb-810c-421130e8d524/pull/0.log" Oct 03 17:52:24 crc kubenswrapper[4744]: I1003 17:52:24.124480 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd_6a7d3bdf-d461-4fcb-810c-421130e8d524/pull/0.log" Oct 03 17:52:24 crc kubenswrapper[4744]: I1003 17:52:24.188182 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd_6a7d3bdf-d461-4fcb-810c-421130e8d524/util/0.log" Oct 03 17:52:24 crc kubenswrapper[4744]: I1003 17:52:24.433458 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd_6a7d3bdf-d461-4fcb-810c-421130e8d524/extract/0.log" Oct 03 17:52:24 crc kubenswrapper[4744]: I1003 17:52:24.466376 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd_6a7d3bdf-d461-4fcb-810c-421130e8d524/pull/0.log" Oct 03 17:52:24 crc kubenswrapper[4744]: I1003 17:52:24.468191 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7v9wq_6d9cbc3a-4ccb-4e1e-8d56-3b457f5eebe2/registry-server/0.log" Oct 03 17:52:24 crc kubenswrapper[4744]: I1003 17:52:24.471140 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835ck6lkd_6a7d3bdf-d461-4fcb-810c-421130e8d524/util/0.log" Oct 03 17:52:24 crc kubenswrapper[4744]: I1003 17:52:24.658376 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xr8s7_bddf7be1-c68d-42ab-853e-153710114dd4/marketplace-operator/0.log" Oct 03 17:52:24 crc kubenswrapper[4744]: I1003 17:52:24.725674 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wmqbq_86654f3a-97fa-456a-b651-93f8ec174725/extract-utilities/0.log" Oct 03 17:52:24 crc kubenswrapper[4744]: I1003 17:52:24.873893 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wmqbq_86654f3a-97fa-456a-b651-93f8ec174725/extract-content/0.log" Oct 03 17:52:24 crc kubenswrapper[4744]: I1003 17:52:24.882665 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wmqbq_86654f3a-97fa-456a-b651-93f8ec174725/extract-content/0.log" Oct 03 17:52:24 crc kubenswrapper[4744]: I1003 17:52:24.885071 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wmqbq_86654f3a-97fa-456a-b651-93f8ec174725/extract-utilities/0.log" Oct 03 17:52:25 crc kubenswrapper[4744]: I1003 17:52:25.025340 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wmqbq_86654f3a-97fa-456a-b651-93f8ec174725/extract-utilities/0.log" Oct 03 17:52:25 crc kubenswrapper[4744]: I1003 17:52:25.054900 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wmqbq_86654f3a-97fa-456a-b651-93f8ec174725/extract-content/0.log" Oct 03 17:52:25 crc kubenswrapper[4744]: I1003 17:52:25.212309 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l8gwc_83f32f6a-6664-4e8f-8676-cefcc31b6e6c/extract-utilities/0.log" Oct 03 17:52:25 crc kubenswrapper[4744]: I1003 17:52:25.248552 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wmqbq_86654f3a-97fa-456a-b651-93f8ec174725/registry-server/0.log" Oct 03 17:52:25 crc kubenswrapper[4744]: I1003 17:52:25.440778 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l8gwc_83f32f6a-6664-4e8f-8676-cefcc31b6e6c/extract-utilities/0.log" Oct 03 17:52:25 crc kubenswrapper[4744]: I1003 17:52:25.452287 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l8gwc_83f32f6a-6664-4e8f-8676-cefcc31b6e6c/extract-content/0.log" Oct 03 17:52:25 crc kubenswrapper[4744]: I1003 17:52:25.475704 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l8gwc_83f32f6a-6664-4e8f-8676-cefcc31b6e6c/extract-content/0.log" Oct 03 17:52:25 crc kubenswrapper[4744]: I1003 17:52:25.611219 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l8gwc_83f32f6a-6664-4e8f-8676-cefcc31b6e6c/extract-utilities/0.log" Oct 03 17:52:25 crc kubenswrapper[4744]: I1003 17:52:25.665273 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l8gwc_83f32f6a-6664-4e8f-8676-cefcc31b6e6c/extract-content/0.log" Oct 03 17:52:26 crc kubenswrapper[4744]: I1003 17:52:26.240680 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-l8gwc_83f32f6a-6664-4e8f-8676-cefcc31b6e6c/registry-server/0.log" Oct 03 17:52:32 crc kubenswrapper[4744]: I1003 17:52:32.891831 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:52:32 crc kubenswrapper[4744]: E1003 17:52:32.892717 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:52:43 crc kubenswrapper[4744]: I1003 17:52:43.892590 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:52:43 crc kubenswrapper[4744]: E1003 17:52:43.893781 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:52:55 crc kubenswrapper[4744]: E1003 17:52:55.255266 4744 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.132:41560->38.102.83.132:38347: read tcp 38.102.83.132:41560->38.102.83.132:38347: read: connection reset by peer Oct 03 17:52:55 crc kubenswrapper[4744]: I1003 17:52:55.969629 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-62vjq"] Oct 03 17:52:55 crc kubenswrapper[4744]: E1003 17:52:55.971183 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270443ae-cab5-4bae-848f-76bcdb3fb610" containerName="container-00" Oct 03 17:52:55 crc kubenswrapper[4744]: I1003 17:52:55.971198 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="270443ae-cab5-4bae-848f-76bcdb3fb610" containerName="container-00" Oct 03 17:52:55 crc kubenswrapper[4744]: I1003 17:52:55.973535 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="270443ae-cab5-4bae-848f-76bcdb3fb610" containerName="container-00" Oct 03 17:52:55 crc kubenswrapper[4744]: I1003 17:52:55.997299 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:52:56 crc kubenswrapper[4744]: I1003 17:52:56.016172 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-62vjq"] Oct 03 17:52:56 crc kubenswrapper[4744]: I1003 17:52:56.185960 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hdjp\" (UniqueName: \"kubernetes.io/projected/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-kube-api-access-9hdjp\") pod \"community-operators-62vjq\" (UID: \"656cab8e-d6fb-4196-b6c5-d19ed872d2bc\") " pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:52:56 crc kubenswrapper[4744]: I1003 17:52:56.186058 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-catalog-content\") pod \"community-operators-62vjq\" (UID: \"656cab8e-d6fb-4196-b6c5-d19ed872d2bc\") " pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:52:56 crc kubenswrapper[4744]: I1003 17:52:56.186141 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-utilities\") pod \"community-operators-62vjq\" (UID: \"656cab8e-d6fb-4196-b6c5-d19ed872d2bc\") " pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:52:56 crc kubenswrapper[4744]: I1003 17:52:56.287739 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hdjp\" (UniqueName: \"kubernetes.io/projected/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-kube-api-access-9hdjp\") pod \"community-operators-62vjq\" (UID: \"656cab8e-d6fb-4196-b6c5-d19ed872d2bc\") " pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:52:56 crc kubenswrapper[4744]: I1003 17:52:56.287849 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-catalog-content\") pod \"community-operators-62vjq\" (UID: \"656cab8e-d6fb-4196-b6c5-d19ed872d2bc\") " pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:52:56 crc kubenswrapper[4744]: I1003 17:52:56.287928 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-utilities\") pod \"community-operators-62vjq\" (UID: \"656cab8e-d6fb-4196-b6c5-d19ed872d2bc\") " pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:52:56 crc kubenswrapper[4744]: I1003 17:52:56.288596 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-catalog-content\") pod \"community-operators-62vjq\" (UID: \"656cab8e-d6fb-4196-b6c5-d19ed872d2bc\") " pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:52:56 crc kubenswrapper[4744]: I1003 17:52:56.288604 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-utilities\") pod \"community-operators-62vjq\" (UID: \"656cab8e-d6fb-4196-b6c5-d19ed872d2bc\") " pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:52:56 crc kubenswrapper[4744]: I1003 17:52:56.320298 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hdjp\" (UniqueName: \"kubernetes.io/projected/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-kube-api-access-9hdjp\") pod \"community-operators-62vjq\" (UID: \"656cab8e-d6fb-4196-b6c5-d19ed872d2bc\") " pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:52:56 crc kubenswrapper[4744]: I1003 17:52:56.335443 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:52:56 crc kubenswrapper[4744]: I1003 17:52:56.967626 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-62vjq"] Oct 03 17:52:57 crc kubenswrapper[4744]: I1003 17:52:57.057847 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62vjq" event={"ID":"656cab8e-d6fb-4196-b6c5-d19ed872d2bc","Type":"ContainerStarted","Data":"56f559b91a30c6922ef5cd589bd571a8e135e93c395cb2f5629d0cf4459096e7"} Oct 03 17:52:57 crc kubenswrapper[4744]: E1003 17:52:57.475588 4744 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.132:46652->38.102.83.132:38347: write tcp 38.102.83.132:46652->38.102.83.132:38347: write: broken pipe Oct 03 17:52:57 crc kubenswrapper[4744]: I1003 17:52:57.892949 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:52:57 crc kubenswrapper[4744]: E1003 17:52:57.893398 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:52:58 crc kubenswrapper[4744]: I1003 17:52:58.066381 4744 generic.go:334] "Generic (PLEG): container finished" podID="656cab8e-d6fb-4196-b6c5-d19ed872d2bc" containerID="8673870bd7fa7a1a0e27265f55f60e782a505e499c6534207af7c3db58ee3f1d" exitCode=0 Oct 03 17:52:58 crc kubenswrapper[4744]: I1003 17:52:58.066419 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62vjq" event={"ID":"656cab8e-d6fb-4196-b6c5-d19ed872d2bc","Type":"ContainerDied","Data":"8673870bd7fa7a1a0e27265f55f60e782a505e499c6534207af7c3db58ee3f1d"} Oct 03 17:53:00 crc kubenswrapper[4744]: I1003 17:53:00.085432 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62vjq" event={"ID":"656cab8e-d6fb-4196-b6c5-d19ed872d2bc","Type":"ContainerStarted","Data":"f236a03299204e9647196c4aeabb97b40034eac04ac5b0f2bf3ab1067e3b7f67"} Oct 03 17:53:01 crc kubenswrapper[4744]: I1003 17:53:01.101901 4744 generic.go:334] "Generic (PLEG): container finished" podID="656cab8e-d6fb-4196-b6c5-d19ed872d2bc" containerID="f236a03299204e9647196c4aeabb97b40034eac04ac5b0f2bf3ab1067e3b7f67" exitCode=0 Oct 03 17:53:01 crc kubenswrapper[4744]: I1003 17:53:01.101984 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62vjq" event={"ID":"656cab8e-d6fb-4196-b6c5-d19ed872d2bc","Type":"ContainerDied","Data":"f236a03299204e9647196c4aeabb97b40034eac04ac5b0f2bf3ab1067e3b7f67"} Oct 03 17:53:02 crc kubenswrapper[4744]: E1003 17:53:02.969037 4744 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.132:46876->38.102.83.132:38347: write tcp 38.102.83.132:46876->38.102.83.132:38347: write: broken pipe Oct 03 17:53:03 crc kubenswrapper[4744]: I1003 17:53:03.125048 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62vjq" event={"ID":"656cab8e-d6fb-4196-b6c5-d19ed872d2bc","Type":"ContainerStarted","Data":"27f1dc425a82b3e651dabeb51018c7b03ae7385d723b777418e9cabeab938880"} Oct 03 17:53:03 crc kubenswrapper[4744]: I1003 17:53:03.165247 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-62vjq" podStartSLOduration=4.627271522 podStartE2EDuration="8.165221572s" podCreationTimestamp="2025-10-03 17:52:55 +0000 UTC" firstStartedPulling="2025-10-03 17:52:58.067875799 +0000 UTC m=+5304.347751695" lastFinishedPulling="2025-10-03 17:53:01.605825849 +0000 UTC m=+5307.885701745" observedRunningTime="2025-10-03 17:53:03.148359404 +0000 UTC m=+5309.428235300" watchObservedRunningTime="2025-10-03 17:53:03.165221572 +0000 UTC m=+5309.445097468" Oct 03 17:53:06 crc kubenswrapper[4744]: I1003 17:53:06.337449 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:53:06 crc kubenswrapper[4744]: I1003 17:53:06.338079 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:53:06 crc kubenswrapper[4744]: I1003 17:53:06.404741 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:53:07 crc kubenswrapper[4744]: I1003 17:53:07.248273 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:53:07 crc kubenswrapper[4744]: I1003 17:53:07.313167 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-62vjq"] Oct 03 17:53:09 crc kubenswrapper[4744]: I1003 17:53:09.196074 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-62vjq" podUID="656cab8e-d6fb-4196-b6c5-d19ed872d2bc" containerName="registry-server" containerID="cri-o://27f1dc425a82b3e651dabeb51018c7b03ae7385d723b777418e9cabeab938880" gracePeriod=2 Oct 03 17:53:09 crc kubenswrapper[4744]: I1003 17:53:09.712886 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:53:09 crc kubenswrapper[4744]: I1003 17:53:09.891973 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:53:09 crc kubenswrapper[4744]: E1003 17:53:09.892276 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:53:09 crc kubenswrapper[4744]: I1003 17:53:09.900544 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hdjp\" (UniqueName: \"kubernetes.io/projected/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-kube-api-access-9hdjp\") pod \"656cab8e-d6fb-4196-b6c5-d19ed872d2bc\" (UID: \"656cab8e-d6fb-4196-b6c5-d19ed872d2bc\") " Oct 03 17:53:09 crc kubenswrapper[4744]: I1003 17:53:09.901738 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-catalog-content\") pod \"656cab8e-d6fb-4196-b6c5-d19ed872d2bc\" (UID: \"656cab8e-d6fb-4196-b6c5-d19ed872d2bc\") " Oct 03 17:53:09 crc kubenswrapper[4744]: I1003 17:53:09.901974 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-utilities\") pod \"656cab8e-d6fb-4196-b6c5-d19ed872d2bc\" (UID: \"656cab8e-d6fb-4196-b6c5-d19ed872d2bc\") " Oct 03 17:53:09 crc kubenswrapper[4744]: I1003 17:53:09.903952 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-utilities" (OuterVolumeSpecName: "utilities") pod "656cab8e-d6fb-4196-b6c5-d19ed872d2bc" (UID: "656cab8e-d6fb-4196-b6c5-d19ed872d2bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:53:09 crc kubenswrapper[4744]: I1003 17:53:09.909690 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-kube-api-access-9hdjp" (OuterVolumeSpecName: "kube-api-access-9hdjp") pod "656cab8e-d6fb-4196-b6c5-d19ed872d2bc" (UID: "656cab8e-d6fb-4196-b6c5-d19ed872d2bc"). InnerVolumeSpecName "kube-api-access-9hdjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:53:09 crc kubenswrapper[4744]: I1003 17:53:09.961063 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "656cab8e-d6fb-4196-b6c5-d19ed872d2bc" (UID: "656cab8e-d6fb-4196-b6c5-d19ed872d2bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.004814 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.004852 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hdjp\" (UniqueName: \"kubernetes.io/projected/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-kube-api-access-9hdjp\") on node \"crc\" DevicePath \"\"" Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.004862 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/656cab8e-d6fb-4196-b6c5-d19ed872d2bc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.204284 4744 generic.go:334] "Generic (PLEG): container finished" podID="656cab8e-d6fb-4196-b6c5-d19ed872d2bc" containerID="27f1dc425a82b3e651dabeb51018c7b03ae7385d723b777418e9cabeab938880" exitCode=0 Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.204613 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62vjq" Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.210775 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62vjq" event={"ID":"656cab8e-d6fb-4196-b6c5-d19ed872d2bc","Type":"ContainerDied","Data":"27f1dc425a82b3e651dabeb51018c7b03ae7385d723b777418e9cabeab938880"} Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.210831 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62vjq" event={"ID":"656cab8e-d6fb-4196-b6c5-d19ed872d2bc","Type":"ContainerDied","Data":"56f559b91a30c6922ef5cd589bd571a8e135e93c395cb2f5629d0cf4459096e7"} Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.210858 4744 scope.go:117] "RemoveContainer" containerID="27f1dc425a82b3e651dabeb51018c7b03ae7385d723b777418e9cabeab938880" Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.239981 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-62vjq"] Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.241065 4744 scope.go:117] "RemoveContainer" containerID="f236a03299204e9647196c4aeabb97b40034eac04ac5b0f2bf3ab1067e3b7f67" Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.254715 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-62vjq"] Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.263905 4744 scope.go:117] "RemoveContainer" containerID="8673870bd7fa7a1a0e27265f55f60e782a505e499c6534207af7c3db58ee3f1d" Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.303666 4744 scope.go:117] "RemoveContainer" containerID="27f1dc425a82b3e651dabeb51018c7b03ae7385d723b777418e9cabeab938880" Oct 03 17:53:10 crc kubenswrapper[4744]: E1003 17:53:10.304159 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27f1dc425a82b3e651dabeb51018c7b03ae7385d723b777418e9cabeab938880\": container with ID starting with 27f1dc425a82b3e651dabeb51018c7b03ae7385d723b777418e9cabeab938880 not found: ID does not exist" containerID="27f1dc425a82b3e651dabeb51018c7b03ae7385d723b777418e9cabeab938880" Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.304187 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27f1dc425a82b3e651dabeb51018c7b03ae7385d723b777418e9cabeab938880"} err="failed to get container status \"27f1dc425a82b3e651dabeb51018c7b03ae7385d723b777418e9cabeab938880\": rpc error: code = NotFound desc = could not find container \"27f1dc425a82b3e651dabeb51018c7b03ae7385d723b777418e9cabeab938880\": container with ID starting with 27f1dc425a82b3e651dabeb51018c7b03ae7385d723b777418e9cabeab938880 not found: ID does not exist" Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.304207 4744 scope.go:117] "RemoveContainer" containerID="f236a03299204e9647196c4aeabb97b40034eac04ac5b0f2bf3ab1067e3b7f67" Oct 03 17:53:10 crc kubenswrapper[4744]: E1003 17:53:10.304444 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f236a03299204e9647196c4aeabb97b40034eac04ac5b0f2bf3ab1067e3b7f67\": container with ID starting with f236a03299204e9647196c4aeabb97b40034eac04ac5b0f2bf3ab1067e3b7f67 not found: ID does not exist" containerID="f236a03299204e9647196c4aeabb97b40034eac04ac5b0f2bf3ab1067e3b7f67" Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.304465 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f236a03299204e9647196c4aeabb97b40034eac04ac5b0f2bf3ab1067e3b7f67"} err="failed to get container status \"f236a03299204e9647196c4aeabb97b40034eac04ac5b0f2bf3ab1067e3b7f67\": rpc error: code = NotFound desc = could not find container \"f236a03299204e9647196c4aeabb97b40034eac04ac5b0f2bf3ab1067e3b7f67\": container with ID starting with f236a03299204e9647196c4aeabb97b40034eac04ac5b0f2bf3ab1067e3b7f67 not found: ID does not exist" Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.304479 4744 scope.go:117] "RemoveContainer" containerID="8673870bd7fa7a1a0e27265f55f60e782a505e499c6534207af7c3db58ee3f1d" Oct 03 17:53:10 crc kubenswrapper[4744]: E1003 17:53:10.304827 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8673870bd7fa7a1a0e27265f55f60e782a505e499c6534207af7c3db58ee3f1d\": container with ID starting with 8673870bd7fa7a1a0e27265f55f60e782a505e499c6534207af7c3db58ee3f1d not found: ID does not exist" containerID="8673870bd7fa7a1a0e27265f55f60e782a505e499c6534207af7c3db58ee3f1d" Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.304850 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8673870bd7fa7a1a0e27265f55f60e782a505e499c6534207af7c3db58ee3f1d"} err="failed to get container status \"8673870bd7fa7a1a0e27265f55f60e782a505e499c6534207af7c3db58ee3f1d\": rpc error: code = NotFound desc = could not find container \"8673870bd7fa7a1a0e27265f55f60e782a505e499c6534207af7c3db58ee3f1d\": container with ID starting with 8673870bd7fa7a1a0e27265f55f60e782a505e499c6534207af7c3db58ee3f1d not found: ID does not exist" Oct 03 17:53:10 crc kubenswrapper[4744]: I1003 17:53:10.908644 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="656cab8e-d6fb-4196-b6c5-d19ed872d2bc" path="/var/lib/kubelet/pods/656cab8e-d6fb-4196-b6c5-d19ed872d2bc/volumes" Oct 03 17:53:22 crc kubenswrapper[4744]: I1003 17:53:22.893395 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:53:22 crc kubenswrapper[4744]: E1003 17:53:22.894548 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:53:36 crc kubenswrapper[4744]: I1003 17:53:36.892973 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:53:36 crc kubenswrapper[4744]: E1003 17:53:36.894193 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:53:49 crc kubenswrapper[4744]: I1003 17:53:49.893365 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:53:49 crc kubenswrapper[4744]: E1003 17:53:49.894114 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:54:00 crc kubenswrapper[4744]: I1003 17:54:00.892415 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:54:00 crc kubenswrapper[4744]: E1003 17:54:00.893234 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:54:12 crc kubenswrapper[4744]: I1003 17:54:12.894336 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:54:12 crc kubenswrapper[4744]: E1003 17:54:12.896721 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:54:24 crc kubenswrapper[4744]: I1003 17:54:24.892582 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:54:24 crc kubenswrapper[4744]: E1003 17:54:24.893927 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.176095 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tfl5z"] Oct 03 17:54:28 crc kubenswrapper[4744]: E1003 17:54:28.177819 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="656cab8e-d6fb-4196-b6c5-d19ed872d2bc" containerName="extract-utilities" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.177898 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="656cab8e-d6fb-4196-b6c5-d19ed872d2bc" containerName="extract-utilities" Oct 03 17:54:28 crc kubenswrapper[4744]: E1003 17:54:28.177981 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="656cab8e-d6fb-4196-b6c5-d19ed872d2bc" containerName="extract-content" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.178055 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="656cab8e-d6fb-4196-b6c5-d19ed872d2bc" containerName="extract-content" Oct 03 17:54:28 crc kubenswrapper[4744]: E1003 17:54:28.178138 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="656cab8e-d6fb-4196-b6c5-d19ed872d2bc" containerName="registry-server" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.178193 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="656cab8e-d6fb-4196-b6c5-d19ed872d2bc" containerName="registry-server" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.178439 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="656cab8e-d6fb-4196-b6c5-d19ed872d2bc" containerName="registry-server" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.179876 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.191386 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tfl5z"] Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.303901 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/848a395e-8062-46fd-b80c-c60ccf2a22e8-utilities\") pod \"redhat-marketplace-tfl5z\" (UID: \"848a395e-8062-46fd-b80c-c60ccf2a22e8\") " pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.303964 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/848a395e-8062-46fd-b80c-c60ccf2a22e8-catalog-content\") pod \"redhat-marketplace-tfl5z\" (UID: \"848a395e-8062-46fd-b80c-c60ccf2a22e8\") " pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.304117 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2gj6\" (UniqueName: \"kubernetes.io/projected/848a395e-8062-46fd-b80c-c60ccf2a22e8-kube-api-access-s2gj6\") pod \"redhat-marketplace-tfl5z\" (UID: \"848a395e-8062-46fd-b80c-c60ccf2a22e8\") " pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.405818 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2gj6\" (UniqueName: \"kubernetes.io/projected/848a395e-8062-46fd-b80c-c60ccf2a22e8-kube-api-access-s2gj6\") pod \"redhat-marketplace-tfl5z\" (UID: \"848a395e-8062-46fd-b80c-c60ccf2a22e8\") " pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.405930 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/848a395e-8062-46fd-b80c-c60ccf2a22e8-utilities\") pod \"redhat-marketplace-tfl5z\" (UID: \"848a395e-8062-46fd-b80c-c60ccf2a22e8\") " pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.405960 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/848a395e-8062-46fd-b80c-c60ccf2a22e8-catalog-content\") pod \"redhat-marketplace-tfl5z\" (UID: \"848a395e-8062-46fd-b80c-c60ccf2a22e8\") " pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.406397 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/848a395e-8062-46fd-b80c-c60ccf2a22e8-catalog-content\") pod \"redhat-marketplace-tfl5z\" (UID: \"848a395e-8062-46fd-b80c-c60ccf2a22e8\") " pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.406861 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/848a395e-8062-46fd-b80c-c60ccf2a22e8-utilities\") pod \"redhat-marketplace-tfl5z\" (UID: \"848a395e-8062-46fd-b80c-c60ccf2a22e8\") " pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.428931 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2gj6\" (UniqueName: \"kubernetes.io/projected/848a395e-8062-46fd-b80c-c60ccf2a22e8-kube-api-access-s2gj6\") pod \"redhat-marketplace-tfl5z\" (UID: \"848a395e-8062-46fd-b80c-c60ccf2a22e8\") " pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.504630 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:28 crc kubenswrapper[4744]: I1003 17:54:28.968462 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tfl5z"] Oct 03 17:54:29 crc kubenswrapper[4744]: I1003 17:54:29.162843 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfl5z" event={"ID":"848a395e-8062-46fd-b80c-c60ccf2a22e8","Type":"ContainerStarted","Data":"bcdc71f517327800724d2b00d52266871fc353cb038b3717968c0c31d3fc2ad3"} Oct 03 17:54:30 crc kubenswrapper[4744]: I1003 17:54:30.174606 4744 generic.go:334] "Generic (PLEG): container finished" podID="848a395e-8062-46fd-b80c-c60ccf2a22e8" containerID="732f8929a84cb3789283c8186a681abd8b974153dc74ae390cbd63d4be05562a" exitCode=0 Oct 03 17:54:30 crc kubenswrapper[4744]: I1003 17:54:30.175016 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfl5z" event={"ID":"848a395e-8062-46fd-b80c-c60ccf2a22e8","Type":"ContainerDied","Data":"732f8929a84cb3789283c8186a681abd8b974153dc74ae390cbd63d4be05562a"} Oct 03 17:54:32 crc kubenswrapper[4744]: I1003 17:54:32.217540 4744 generic.go:334] "Generic (PLEG): container finished" podID="848a395e-8062-46fd-b80c-c60ccf2a22e8" containerID="d7be443701c6fde93cb2689c1d2567532d3c34d4074b8d4e42590b089080abd1" exitCode=0 Oct 03 17:54:32 crc kubenswrapper[4744]: I1003 17:54:32.217645 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfl5z" event={"ID":"848a395e-8062-46fd-b80c-c60ccf2a22e8","Type":"ContainerDied","Data":"d7be443701c6fde93cb2689c1d2567532d3c34d4074b8d4e42590b089080abd1"} Oct 03 17:54:33 crc kubenswrapper[4744]: I1003 17:54:33.230064 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfl5z" event={"ID":"848a395e-8062-46fd-b80c-c60ccf2a22e8","Type":"ContainerStarted","Data":"99b9ad7c364bdfb907d699cf97a30e8674b43bbd43b832953c239603a1842c98"} Oct 03 17:54:33 crc kubenswrapper[4744]: I1003 17:54:33.266358 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tfl5z" podStartSLOduration=2.806276312 podStartE2EDuration="5.266331071s" podCreationTimestamp="2025-10-03 17:54:28 +0000 UTC" firstStartedPulling="2025-10-03 17:54:30.176929235 +0000 UTC m=+5396.456805131" lastFinishedPulling="2025-10-03 17:54:32.636983954 +0000 UTC m=+5398.916859890" observedRunningTime="2025-10-03 17:54:33.248912969 +0000 UTC m=+5399.528788915" watchObservedRunningTime="2025-10-03 17:54:33.266331071 +0000 UTC m=+5399.546207007" Oct 03 17:54:37 crc kubenswrapper[4744]: I1003 17:54:37.896806 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:54:37 crc kubenswrapper[4744]: E1003 17:54:37.897810 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:54:38 crc kubenswrapper[4744]: I1003 17:54:38.505593 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:38 crc kubenswrapper[4744]: I1003 17:54:38.507223 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:38 crc kubenswrapper[4744]: I1003 17:54:38.602873 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:39 crc kubenswrapper[4744]: I1003 17:54:39.376715 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:39 crc kubenswrapper[4744]: I1003 17:54:39.465288 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tfl5z"] Oct 03 17:54:41 crc kubenswrapper[4744]: I1003 17:54:41.332004 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tfl5z" podUID="848a395e-8062-46fd-b80c-c60ccf2a22e8" containerName="registry-server" containerID="cri-o://99b9ad7c364bdfb907d699cf97a30e8674b43bbd43b832953c239603a1842c98" gracePeriod=2 Oct 03 17:54:41 crc kubenswrapper[4744]: I1003 17:54:41.835978 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:41 crc kubenswrapper[4744]: I1003 17:54:41.948736 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2gj6\" (UniqueName: \"kubernetes.io/projected/848a395e-8062-46fd-b80c-c60ccf2a22e8-kube-api-access-s2gj6\") pod \"848a395e-8062-46fd-b80c-c60ccf2a22e8\" (UID: \"848a395e-8062-46fd-b80c-c60ccf2a22e8\") " Oct 03 17:54:41 crc kubenswrapper[4744]: I1003 17:54:41.948840 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/848a395e-8062-46fd-b80c-c60ccf2a22e8-utilities\") pod \"848a395e-8062-46fd-b80c-c60ccf2a22e8\" (UID: \"848a395e-8062-46fd-b80c-c60ccf2a22e8\") " Oct 03 17:54:41 crc kubenswrapper[4744]: I1003 17:54:41.948886 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/848a395e-8062-46fd-b80c-c60ccf2a22e8-catalog-content\") pod \"848a395e-8062-46fd-b80c-c60ccf2a22e8\" (UID: \"848a395e-8062-46fd-b80c-c60ccf2a22e8\") " Oct 03 17:54:41 crc kubenswrapper[4744]: I1003 17:54:41.949906 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/848a395e-8062-46fd-b80c-c60ccf2a22e8-utilities" (OuterVolumeSpecName: "utilities") pod "848a395e-8062-46fd-b80c-c60ccf2a22e8" (UID: "848a395e-8062-46fd-b80c-c60ccf2a22e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:54:41 crc kubenswrapper[4744]: I1003 17:54:41.954863 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/848a395e-8062-46fd-b80c-c60ccf2a22e8-kube-api-access-s2gj6" (OuterVolumeSpecName: "kube-api-access-s2gj6") pod "848a395e-8062-46fd-b80c-c60ccf2a22e8" (UID: "848a395e-8062-46fd-b80c-c60ccf2a22e8"). InnerVolumeSpecName "kube-api-access-s2gj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:54:41 crc kubenswrapper[4744]: I1003 17:54:41.990558 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/848a395e-8062-46fd-b80c-c60ccf2a22e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "848a395e-8062-46fd-b80c-c60ccf2a22e8" (UID: "848a395e-8062-46fd-b80c-c60ccf2a22e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.050679 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2gj6\" (UniqueName: \"kubernetes.io/projected/848a395e-8062-46fd-b80c-c60ccf2a22e8-kube-api-access-s2gj6\") on node \"crc\" DevicePath \"\"" Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.050716 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/848a395e-8062-46fd-b80c-c60ccf2a22e8-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.050731 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/848a395e-8062-46fd-b80c-c60ccf2a22e8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.349552 4744 generic.go:334] "Generic (PLEG): container finished" podID="848a395e-8062-46fd-b80c-c60ccf2a22e8" containerID="99b9ad7c364bdfb907d699cf97a30e8674b43bbd43b832953c239603a1842c98" exitCode=0 Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.349608 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfl5z" event={"ID":"848a395e-8062-46fd-b80c-c60ccf2a22e8","Type":"ContainerDied","Data":"99b9ad7c364bdfb907d699cf97a30e8674b43bbd43b832953c239603a1842c98"} Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.349649 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tfl5z" Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.349701 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfl5z" event={"ID":"848a395e-8062-46fd-b80c-c60ccf2a22e8","Type":"ContainerDied","Data":"bcdc71f517327800724d2b00d52266871fc353cb038b3717968c0c31d3fc2ad3"} Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.349768 4744 scope.go:117] "RemoveContainer" containerID="99b9ad7c364bdfb907d699cf97a30e8674b43bbd43b832953c239603a1842c98" Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.387713 4744 scope.go:117] "RemoveContainer" containerID="d7be443701c6fde93cb2689c1d2567532d3c34d4074b8d4e42590b089080abd1" Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.402832 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tfl5z"] Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.416701 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tfl5z"] Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.444744 4744 scope.go:117] "RemoveContainer" containerID="732f8929a84cb3789283c8186a681abd8b974153dc74ae390cbd63d4be05562a" Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.487984 4744 scope.go:117] "RemoveContainer" containerID="99b9ad7c364bdfb907d699cf97a30e8674b43bbd43b832953c239603a1842c98" Oct 03 17:54:42 crc kubenswrapper[4744]: E1003 17:54:42.488524 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99b9ad7c364bdfb907d699cf97a30e8674b43bbd43b832953c239603a1842c98\": container with ID starting with 99b9ad7c364bdfb907d699cf97a30e8674b43bbd43b832953c239603a1842c98 not found: ID does not exist" containerID="99b9ad7c364bdfb907d699cf97a30e8674b43bbd43b832953c239603a1842c98" Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.488596 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99b9ad7c364bdfb907d699cf97a30e8674b43bbd43b832953c239603a1842c98"} err="failed to get container status \"99b9ad7c364bdfb907d699cf97a30e8674b43bbd43b832953c239603a1842c98\": rpc error: code = NotFound desc = could not find container \"99b9ad7c364bdfb907d699cf97a30e8674b43bbd43b832953c239603a1842c98\": container with ID starting with 99b9ad7c364bdfb907d699cf97a30e8674b43bbd43b832953c239603a1842c98 not found: ID does not exist" Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.488637 4744 scope.go:117] "RemoveContainer" containerID="d7be443701c6fde93cb2689c1d2567532d3c34d4074b8d4e42590b089080abd1" Oct 03 17:54:42 crc kubenswrapper[4744]: E1003 17:54:42.489290 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7be443701c6fde93cb2689c1d2567532d3c34d4074b8d4e42590b089080abd1\": container with ID starting with d7be443701c6fde93cb2689c1d2567532d3c34d4074b8d4e42590b089080abd1 not found: ID does not exist" containerID="d7be443701c6fde93cb2689c1d2567532d3c34d4074b8d4e42590b089080abd1" Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.489343 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7be443701c6fde93cb2689c1d2567532d3c34d4074b8d4e42590b089080abd1"} err="failed to get container status \"d7be443701c6fde93cb2689c1d2567532d3c34d4074b8d4e42590b089080abd1\": rpc error: code = NotFound desc = could not find container \"d7be443701c6fde93cb2689c1d2567532d3c34d4074b8d4e42590b089080abd1\": container with ID starting with d7be443701c6fde93cb2689c1d2567532d3c34d4074b8d4e42590b089080abd1 not found: ID does not exist" Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.489374 4744 scope.go:117] "RemoveContainer" containerID="732f8929a84cb3789283c8186a681abd8b974153dc74ae390cbd63d4be05562a" Oct 03 17:54:42 crc kubenswrapper[4744]: E1003 17:54:42.489751 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"732f8929a84cb3789283c8186a681abd8b974153dc74ae390cbd63d4be05562a\": container with ID starting with 732f8929a84cb3789283c8186a681abd8b974153dc74ae390cbd63d4be05562a not found: ID does not exist" containerID="732f8929a84cb3789283c8186a681abd8b974153dc74ae390cbd63d4be05562a" Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.489791 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"732f8929a84cb3789283c8186a681abd8b974153dc74ae390cbd63d4be05562a"} err="failed to get container status \"732f8929a84cb3789283c8186a681abd8b974153dc74ae390cbd63d4be05562a\": rpc error: code = NotFound desc = could not find container \"732f8929a84cb3789283c8186a681abd8b974153dc74ae390cbd63d4be05562a\": container with ID starting with 732f8929a84cb3789283c8186a681abd8b974153dc74ae390cbd63d4be05562a not found: ID does not exist" Oct 03 17:54:42 crc kubenswrapper[4744]: I1003 17:54:42.912921 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="848a395e-8062-46fd-b80c-c60ccf2a22e8" path="/var/lib/kubelet/pods/848a395e-8062-46fd-b80c-c60ccf2a22e8/volumes" Oct 03 17:54:49 crc kubenswrapper[4744]: I1003 17:54:49.891804 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:54:49 crc kubenswrapper[4744]: E1003 17:54:49.893071 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:54:53 crc kubenswrapper[4744]: I1003 17:54:53.517083 4744 generic.go:334] "Generic (PLEG): container finished" podID="3e32a7a9-b8ca-4900-a343-22c70c44f69c" containerID="3e7292228fd4f55ef95277d281f282b4ec45ebc06d8ddfca00fbf108c4be6cdd" exitCode=0 Oct 03 17:54:53 crc kubenswrapper[4744]: I1003 17:54:53.517197 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mgmdg/must-gather-5h7jg" event={"ID":"3e32a7a9-b8ca-4900-a343-22c70c44f69c","Type":"ContainerDied","Data":"3e7292228fd4f55ef95277d281f282b4ec45ebc06d8ddfca00fbf108c4be6cdd"} Oct 03 17:54:53 crc kubenswrapper[4744]: I1003 17:54:53.518926 4744 scope.go:117] "RemoveContainer" containerID="3e7292228fd4f55ef95277d281f282b4ec45ebc06d8ddfca00fbf108c4be6cdd" Oct 03 17:54:54 crc kubenswrapper[4744]: I1003 17:54:54.074778 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mgmdg_must-gather-5h7jg_3e32a7a9-b8ca-4900-a343-22c70c44f69c/gather/0.log" Oct 03 17:55:00 crc kubenswrapper[4744]: I1003 17:55:00.892662 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:55:00 crc kubenswrapper[4744]: E1003 17:55:00.893394 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.178912 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mgmdg/must-gather-5h7jg"] Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.179677 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mgmdg/must-gather-5h7jg" podUID="3e32a7a9-b8ca-4900-a343-22c70c44f69c" containerName="copy" containerID="cri-o://ab51547dffcfedac9a57e4c70d3ccc11d31f6b850f608e1635db38e434486693" gracePeriod=2 Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.190790 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mgmdg/must-gather-5h7jg"] Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.616061 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mgmdg_must-gather-5h7jg_3e32a7a9-b8ca-4900-a343-22c70c44f69c/copy/0.log" Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.616761 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/must-gather-5h7jg" Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.687015 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mgmdg_must-gather-5h7jg_3e32a7a9-b8ca-4900-a343-22c70c44f69c/copy/0.log" Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.687509 4744 generic.go:334] "Generic (PLEG): container finished" podID="3e32a7a9-b8ca-4900-a343-22c70c44f69c" containerID="ab51547dffcfedac9a57e4c70d3ccc11d31f6b850f608e1635db38e434486693" exitCode=143 Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.687576 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mgmdg/must-gather-5h7jg" Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.687650 4744 scope.go:117] "RemoveContainer" containerID="ab51547dffcfedac9a57e4c70d3ccc11d31f6b850f608e1635db38e434486693" Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.712583 4744 scope.go:117] "RemoveContainer" containerID="3e7292228fd4f55ef95277d281f282b4ec45ebc06d8ddfca00fbf108c4be6cdd" Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.805388 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3e32a7a9-b8ca-4900-a343-22c70c44f69c-must-gather-output\") pod \"3e32a7a9-b8ca-4900-a343-22c70c44f69c\" (UID: \"3e32a7a9-b8ca-4900-a343-22c70c44f69c\") " Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.805514 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sklhb\" (UniqueName: \"kubernetes.io/projected/3e32a7a9-b8ca-4900-a343-22c70c44f69c-kube-api-access-sklhb\") pod \"3e32a7a9-b8ca-4900-a343-22c70c44f69c\" (UID: \"3e32a7a9-b8ca-4900-a343-22c70c44f69c\") " Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.823689 4744 scope.go:117] "RemoveContainer" containerID="ab51547dffcfedac9a57e4c70d3ccc11d31f6b850f608e1635db38e434486693" Oct 03 17:55:07 crc kubenswrapper[4744]: E1003 17:55:07.827760 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab51547dffcfedac9a57e4c70d3ccc11d31f6b850f608e1635db38e434486693\": container with ID starting with ab51547dffcfedac9a57e4c70d3ccc11d31f6b850f608e1635db38e434486693 not found: ID does not exist" containerID="ab51547dffcfedac9a57e4c70d3ccc11d31f6b850f608e1635db38e434486693" Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.827807 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab51547dffcfedac9a57e4c70d3ccc11d31f6b850f608e1635db38e434486693"} err="failed to get container status \"ab51547dffcfedac9a57e4c70d3ccc11d31f6b850f608e1635db38e434486693\": rpc error: code = NotFound desc = could not find container \"ab51547dffcfedac9a57e4c70d3ccc11d31f6b850f608e1635db38e434486693\": container with ID starting with ab51547dffcfedac9a57e4c70d3ccc11d31f6b850f608e1635db38e434486693 not found: ID does not exist" Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.827837 4744 scope.go:117] "RemoveContainer" containerID="3e7292228fd4f55ef95277d281f282b4ec45ebc06d8ddfca00fbf108c4be6cdd" Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.828053 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e32a7a9-b8ca-4900-a343-22c70c44f69c-kube-api-access-sklhb" (OuterVolumeSpecName: "kube-api-access-sklhb") pod "3e32a7a9-b8ca-4900-a343-22c70c44f69c" (UID: "3e32a7a9-b8ca-4900-a343-22c70c44f69c"). InnerVolumeSpecName "kube-api-access-sklhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:55:07 crc kubenswrapper[4744]: E1003 17:55:07.828343 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e7292228fd4f55ef95277d281f282b4ec45ebc06d8ddfca00fbf108c4be6cdd\": container with ID starting with 3e7292228fd4f55ef95277d281f282b4ec45ebc06d8ddfca00fbf108c4be6cdd not found: ID does not exist" containerID="3e7292228fd4f55ef95277d281f282b4ec45ebc06d8ddfca00fbf108c4be6cdd" Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.828372 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e7292228fd4f55ef95277d281f282b4ec45ebc06d8ddfca00fbf108c4be6cdd"} err="failed to get container status \"3e7292228fd4f55ef95277d281f282b4ec45ebc06d8ddfca00fbf108c4be6cdd\": rpc error: code = NotFound desc = could not find container \"3e7292228fd4f55ef95277d281f282b4ec45ebc06d8ddfca00fbf108c4be6cdd\": container with ID starting with 3e7292228fd4f55ef95277d281f282b4ec45ebc06d8ddfca00fbf108c4be6cdd not found: ID does not exist" Oct 03 17:55:07 crc kubenswrapper[4744]: I1003 17:55:07.908179 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sklhb\" (UniqueName: \"kubernetes.io/projected/3e32a7a9-b8ca-4900-a343-22c70c44f69c-kube-api-access-sklhb\") on node \"crc\" DevicePath \"\"" Oct 03 17:55:08 crc kubenswrapper[4744]: I1003 17:55:08.005798 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e32a7a9-b8ca-4900-a343-22c70c44f69c-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3e32a7a9-b8ca-4900-a343-22c70c44f69c" (UID: "3e32a7a9-b8ca-4900-a343-22c70c44f69c"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:55:08 crc kubenswrapper[4744]: I1003 17:55:08.010078 4744 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3e32a7a9-b8ca-4900-a343-22c70c44f69c-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 03 17:55:08 crc kubenswrapper[4744]: E1003 17:55:08.450803 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e32a7a9_b8ca_4900_a343_22c70c44f69c.slice\": RecentStats: unable to find data in memory cache]" Oct 03 17:55:08 crc kubenswrapper[4744]: I1003 17:55:08.921623 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e32a7a9-b8ca-4900-a343-22c70c44f69c" path="/var/lib/kubelet/pods/3e32a7a9-b8ca-4900-a343-22c70c44f69c/volumes" Oct 03 17:55:11 crc kubenswrapper[4744]: I1003 17:55:11.893486 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:55:11 crc kubenswrapper[4744]: E1003 17:55:11.894213 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:55:24 crc kubenswrapper[4744]: I1003 17:55:24.906716 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:55:24 crc kubenswrapper[4744]: E1003 17:55:24.909433 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:55:35 crc kubenswrapper[4744]: I1003 17:55:35.892680 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:55:35 crc kubenswrapper[4744]: E1003 17:55:35.893997 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:55:41 crc kubenswrapper[4744]: I1003 17:55:41.627476 4744 scope.go:117] "RemoveContainer" containerID="7f5e513fdbb52c3368824c4108beb7191433497048df2c33f4b5428f50aa3408" Oct 03 17:55:49 crc kubenswrapper[4744]: I1003 17:55:49.892763 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:55:49 crc kubenswrapper[4744]: E1003 17:55:49.893742 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:56:01 crc kubenswrapper[4744]: I1003 17:56:01.892710 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:56:01 crc kubenswrapper[4744]: E1003 17:56:01.893477 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qzw8f_openshift-machine-config-operator(cc9bea96-4847-44c8-ab50-d469a6c6155d)\"" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" podUID="cc9bea96-4847-44c8-ab50-d469a6c6155d" Oct 03 17:56:12 crc kubenswrapper[4744]: I1003 17:56:12.893172 4744 scope.go:117] "RemoveContainer" containerID="8a9c1b214390bc371953f9703a3dc2a5c84db9cddb5012d7e496b000008d7cf2" Oct 03 17:56:13 crc kubenswrapper[4744]: I1003 17:56:13.429032 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qzw8f" event={"ID":"cc9bea96-4847-44c8-ab50-d469a6c6155d","Type":"ContainerStarted","Data":"dae3e72a92ebb23cb75c935733631ed5afd39a89aad5a27f248dad61c750d535"} Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.703894 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sldp6"] Oct 03 17:56:22 crc kubenswrapper[4744]: E1003 17:56:22.705033 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e32a7a9-b8ca-4900-a343-22c70c44f69c" containerName="gather" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.705053 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e32a7a9-b8ca-4900-a343-22c70c44f69c" containerName="gather" Oct 03 17:56:22 crc kubenswrapper[4744]: E1003 17:56:22.705069 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="848a395e-8062-46fd-b80c-c60ccf2a22e8" containerName="registry-server" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.705079 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="848a395e-8062-46fd-b80c-c60ccf2a22e8" containerName="registry-server" Oct 03 17:56:22 crc kubenswrapper[4744]: E1003 17:56:22.705186 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="848a395e-8062-46fd-b80c-c60ccf2a22e8" containerName="extract-utilities" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.705195 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="848a395e-8062-46fd-b80c-c60ccf2a22e8" containerName="extract-utilities" Oct 03 17:56:22 crc kubenswrapper[4744]: E1003 17:56:22.705216 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e32a7a9-b8ca-4900-a343-22c70c44f69c" containerName="copy" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.705224 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e32a7a9-b8ca-4900-a343-22c70c44f69c" containerName="copy" Oct 03 17:56:22 crc kubenswrapper[4744]: E1003 17:56:22.705253 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="848a395e-8062-46fd-b80c-c60ccf2a22e8" containerName="extract-content" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.705261 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="848a395e-8062-46fd-b80c-c60ccf2a22e8" containerName="extract-content" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.705569 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e32a7a9-b8ca-4900-a343-22c70c44f69c" containerName="copy" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.705597 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e32a7a9-b8ca-4900-a343-22c70c44f69c" containerName="gather" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.705618 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="848a395e-8062-46fd-b80c-c60ccf2a22e8" containerName="registry-server" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.707307 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.721491 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sldp6"] Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.893571 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-catalog-content\") pod \"redhat-operators-sldp6\" (UID: \"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc\") " pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.893718 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-utilities\") pod \"redhat-operators-sldp6\" (UID: \"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc\") " pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.893791 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdvb7\" (UniqueName: \"kubernetes.io/projected/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-kube-api-access-bdvb7\") pod \"redhat-operators-sldp6\" (UID: \"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc\") " pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.995699 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-catalog-content\") pod \"redhat-operators-sldp6\" (UID: \"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc\") " pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.995806 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-utilities\") pod \"redhat-operators-sldp6\" (UID: \"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc\") " pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.995875 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdvb7\" (UniqueName: \"kubernetes.io/projected/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-kube-api-access-bdvb7\") pod \"redhat-operators-sldp6\" (UID: \"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc\") " pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.996277 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-utilities\") pod \"redhat-operators-sldp6\" (UID: \"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc\") " pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:22 crc kubenswrapper[4744]: I1003 17:56:22.996762 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-catalog-content\") pod \"redhat-operators-sldp6\" (UID: \"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc\") " pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:23 crc kubenswrapper[4744]: I1003 17:56:23.020355 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdvb7\" (UniqueName: \"kubernetes.io/projected/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-kube-api-access-bdvb7\") pod \"redhat-operators-sldp6\" (UID: \"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc\") " pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:23 crc kubenswrapper[4744]: I1003 17:56:23.027934 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:23 crc kubenswrapper[4744]: I1003 17:56:23.505060 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sldp6"] Oct 03 17:56:23 crc kubenswrapper[4744]: I1003 17:56:23.537972 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sldp6" event={"ID":"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc","Type":"ContainerStarted","Data":"a63e594bd1ab047911693a95494983d8c32b0fba879f7397be4b7289ac2acd66"} Oct 03 17:56:24 crc kubenswrapper[4744]: I1003 17:56:24.563838 4744 generic.go:334] "Generic (PLEG): container finished" podID="4f89f396-6bae-4e51-9388-2e4c8ac9d6bc" containerID="21d7c64b2e2210b275af6349d96a66aa914dd0e5165fe290d8b75746770baec5" exitCode=0 Oct 03 17:56:24 crc kubenswrapper[4744]: I1003 17:56:24.563884 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sldp6" event={"ID":"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc","Type":"ContainerDied","Data":"21d7c64b2e2210b275af6349d96a66aa914dd0e5165fe290d8b75746770baec5"} Oct 03 17:56:24 crc kubenswrapper[4744]: I1003 17:56:24.567777 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 17:56:26 crc kubenswrapper[4744]: I1003 17:56:26.584950 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sldp6" event={"ID":"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc","Type":"ContainerStarted","Data":"fca05d1eb1030053ef219ef6a4646c98b41775b5cc7979f482902651d039bd6e"} Oct 03 17:56:27 crc kubenswrapper[4744]: I1003 17:56:27.596037 4744 generic.go:334] "Generic (PLEG): container finished" podID="4f89f396-6bae-4e51-9388-2e4c8ac9d6bc" containerID="fca05d1eb1030053ef219ef6a4646c98b41775b5cc7979f482902651d039bd6e" exitCode=0 Oct 03 17:56:27 crc kubenswrapper[4744]: I1003 17:56:27.596086 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sldp6" event={"ID":"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc","Type":"ContainerDied","Data":"fca05d1eb1030053ef219ef6a4646c98b41775b5cc7979f482902651d039bd6e"} Oct 03 17:56:28 crc kubenswrapper[4744]: I1003 17:56:28.607402 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sldp6" event={"ID":"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc","Type":"ContainerStarted","Data":"7249ce9ee4b80b65a177bf968e44915ca52899724d235c65d39a0571ae67ec89"} Oct 03 17:56:28 crc kubenswrapper[4744]: I1003 17:56:28.631010 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sldp6" podStartSLOduration=3.154081963 podStartE2EDuration="6.630989069s" podCreationTimestamp="2025-10-03 17:56:22 +0000 UTC" firstStartedPulling="2025-10-03 17:56:24.567359442 +0000 UTC m=+5510.847235348" lastFinishedPulling="2025-10-03 17:56:28.044266538 +0000 UTC m=+5514.324142454" observedRunningTime="2025-10-03 17:56:28.625372389 +0000 UTC m=+5514.905248275" watchObservedRunningTime="2025-10-03 17:56:28.630989069 +0000 UTC m=+5514.910864965" Oct 03 17:56:33 crc kubenswrapper[4744]: I1003 17:56:33.029011 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:33 crc kubenswrapper[4744]: I1003 17:56:33.029748 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:34 crc kubenswrapper[4744]: I1003 17:56:34.106134 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sldp6" podUID="4f89f396-6bae-4e51-9388-2e4c8ac9d6bc" containerName="registry-server" probeResult="failure" output=< Oct 03 17:56:34 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Oct 03 17:56:34 crc kubenswrapper[4744]: > Oct 03 17:56:43 crc kubenswrapper[4744]: I1003 17:56:43.114738 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:43 crc kubenswrapper[4744]: I1003 17:56:43.199631 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:43 crc kubenswrapper[4744]: I1003 17:56:43.362523 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sldp6"] Oct 03 17:56:44 crc kubenswrapper[4744]: I1003 17:56:44.799829 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sldp6" podUID="4f89f396-6bae-4e51-9388-2e4c8ac9d6bc" containerName="registry-server" containerID="cri-o://7249ce9ee4b80b65a177bf968e44915ca52899724d235c65d39a0571ae67ec89" gracePeriod=2 Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.257124 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.450386 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdvb7\" (UniqueName: \"kubernetes.io/projected/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-kube-api-access-bdvb7\") pod \"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc\" (UID: \"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc\") " Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.450482 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-catalog-content\") pod \"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc\" (UID: \"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc\") " Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.450610 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-utilities\") pod \"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc\" (UID: \"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc\") " Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.452606 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-utilities" (OuterVolumeSpecName: "utilities") pod "4f89f396-6bae-4e51-9388-2e4c8ac9d6bc" (UID: "4f89f396-6bae-4e51-9388-2e4c8ac9d6bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.458726 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-kube-api-access-bdvb7" (OuterVolumeSpecName: "kube-api-access-bdvb7") pod "4f89f396-6bae-4e51-9388-2e4c8ac9d6bc" (UID: "4f89f396-6bae-4e51-9388-2e4c8ac9d6bc"). InnerVolumeSpecName "kube-api-access-bdvb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.553404 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdvb7\" (UniqueName: \"kubernetes.io/projected/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-kube-api-access-bdvb7\") on node \"crc\" DevicePath \"\"" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.553456 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.574348 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4f89f396-6bae-4e51-9388-2e4c8ac9d6bc" (UID: "4f89f396-6bae-4e51-9388-2e4c8ac9d6bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.655927 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.817102 4744 generic.go:334] "Generic (PLEG): container finished" podID="4f89f396-6bae-4e51-9388-2e4c8ac9d6bc" containerID="7249ce9ee4b80b65a177bf968e44915ca52899724d235c65d39a0571ae67ec89" exitCode=0 Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.817140 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sldp6" event={"ID":"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc","Type":"ContainerDied","Data":"7249ce9ee4b80b65a177bf968e44915ca52899724d235c65d39a0571ae67ec89"} Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.817167 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sldp6" event={"ID":"4f89f396-6bae-4e51-9388-2e4c8ac9d6bc","Type":"ContainerDied","Data":"a63e594bd1ab047911693a95494983d8c32b0fba879f7397be4b7289ac2acd66"} Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.817163 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sldp6" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.817182 4744 scope.go:117] "RemoveContainer" containerID="7249ce9ee4b80b65a177bf968e44915ca52899724d235c65d39a0571ae67ec89" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.854263 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sldp6"] Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.862344 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sldp6"] Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.862568 4744 scope.go:117] "RemoveContainer" containerID="fca05d1eb1030053ef219ef6a4646c98b41775b5cc7979f482902651d039bd6e" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.886790 4744 scope.go:117] "RemoveContainer" containerID="21d7c64b2e2210b275af6349d96a66aa914dd0e5165fe290d8b75746770baec5" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.949117 4744 scope.go:117] "RemoveContainer" containerID="7249ce9ee4b80b65a177bf968e44915ca52899724d235c65d39a0571ae67ec89" Oct 03 17:56:45 crc kubenswrapper[4744]: E1003 17:56:45.949937 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7249ce9ee4b80b65a177bf968e44915ca52899724d235c65d39a0571ae67ec89\": container with ID starting with 7249ce9ee4b80b65a177bf968e44915ca52899724d235c65d39a0571ae67ec89 not found: ID does not exist" containerID="7249ce9ee4b80b65a177bf968e44915ca52899724d235c65d39a0571ae67ec89" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.949984 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7249ce9ee4b80b65a177bf968e44915ca52899724d235c65d39a0571ae67ec89"} err="failed to get container status \"7249ce9ee4b80b65a177bf968e44915ca52899724d235c65d39a0571ae67ec89\": rpc error: code = NotFound desc = could not find container \"7249ce9ee4b80b65a177bf968e44915ca52899724d235c65d39a0571ae67ec89\": container with ID starting with 7249ce9ee4b80b65a177bf968e44915ca52899724d235c65d39a0571ae67ec89 not found: ID does not exist" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.950010 4744 scope.go:117] "RemoveContainer" containerID="fca05d1eb1030053ef219ef6a4646c98b41775b5cc7979f482902651d039bd6e" Oct 03 17:56:45 crc kubenswrapper[4744]: E1003 17:56:45.950733 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fca05d1eb1030053ef219ef6a4646c98b41775b5cc7979f482902651d039bd6e\": container with ID starting with fca05d1eb1030053ef219ef6a4646c98b41775b5cc7979f482902651d039bd6e not found: ID does not exist" containerID="fca05d1eb1030053ef219ef6a4646c98b41775b5cc7979f482902651d039bd6e" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.950768 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fca05d1eb1030053ef219ef6a4646c98b41775b5cc7979f482902651d039bd6e"} err="failed to get container status \"fca05d1eb1030053ef219ef6a4646c98b41775b5cc7979f482902651d039bd6e\": rpc error: code = NotFound desc = could not find container \"fca05d1eb1030053ef219ef6a4646c98b41775b5cc7979f482902651d039bd6e\": container with ID starting with fca05d1eb1030053ef219ef6a4646c98b41775b5cc7979f482902651d039bd6e not found: ID does not exist" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.950786 4744 scope.go:117] "RemoveContainer" containerID="21d7c64b2e2210b275af6349d96a66aa914dd0e5165fe290d8b75746770baec5" Oct 03 17:56:45 crc kubenswrapper[4744]: E1003 17:56:45.951176 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21d7c64b2e2210b275af6349d96a66aa914dd0e5165fe290d8b75746770baec5\": container with ID starting with 21d7c64b2e2210b275af6349d96a66aa914dd0e5165fe290d8b75746770baec5 not found: ID does not exist" containerID="21d7c64b2e2210b275af6349d96a66aa914dd0e5165fe290d8b75746770baec5" Oct 03 17:56:45 crc kubenswrapper[4744]: I1003 17:56:45.951206 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21d7c64b2e2210b275af6349d96a66aa914dd0e5165fe290d8b75746770baec5"} err="failed to get container status \"21d7c64b2e2210b275af6349d96a66aa914dd0e5165fe290d8b75746770baec5\": rpc error: code = NotFound desc = could not find container \"21d7c64b2e2210b275af6349d96a66aa914dd0e5165fe290d8b75746770baec5\": container with ID starting with 21d7c64b2e2210b275af6349d96a66aa914dd0e5165fe290d8b75746770baec5 not found: ID does not exist" Oct 03 17:56:46 crc kubenswrapper[4744]: I1003 17:56:46.913883 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f89f396-6bae-4e51-9388-2e4c8ac9d6bc" path="/var/lib/kubelet/pods/4f89f396-6bae-4e51-9388-2e4c8ac9d6bc/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515070007167024450 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015070007170017357 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067773733016527 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067773734015500 5ustar corecore